Skip to main content

Table 1 End-to-End joint training

From: Spatial and temporal learning representation for end-to-end recording device identification

Optimization of end-to-end parameter update algorithm based on deep-and-shallow loss
dataset: 45 device sources, each device source is 514 sentences for training
1: initialized: Wtem, Wspa, Watt, Wclass
2: for k=1,...,K (K=epoch)
3: for t=1,...,T (T=514/batch size)
4: deep-and-shallow loss: L(t)=λ1LT(t)+λ2LS(t)+(1−λ1λ2)LA(t)
5: backpropagation error:\(\frac {\partial L(t)}{\partial x_{i}(t)}\)
6: \(W_{\text {tem}}(t+1){\gets }W_{\text {tem}}(t)-\mu * \frac {\partial L_{T}(t)}{\partial W_{\text {tem}}(t)}-\mu * \frac {\partial L_{A}(t)}{\partial W_{\text {tem}}(t)}\)
7: \(W_{\text {spa}}(t+1){\gets }W_{\text {spa}}(t)-\mu * \frac {\partial L_{s}(t)}{\partial W_{\text {spa}}(t)}-\mu * \frac {\partial L_{A}(t)}{\partial W_{\text {spa}}(t)}\)
8: \(W_{\text {att}}(t+1){\gets }W_{\text {att}}(t)-\mu * \frac {\partial L_{A}(t)}{\partial W_{\text {att}}(t)}\)
9: \(W_{\text {cla}}(t+1){\gets }W_{\text {cla}}(t)-\mu * \frac {\partial L_{A}(t)}{\partial W_{\text {cla}}(t)}\)
10: end for
11: end for
12: return Wtem, Wspa, Watt, Wclass