Full-Capacity Unitary Recurrent Neural Networks

Recurrent neural networks are powerful models for processing sequential data, but they are generally plagued by vanishing and exploding gradient problems. Unitary recurrent neural networks (uRNNs), which use unitary recurrence matrices, have recently…

Authors: Scott Wisdom, Thomas Powers, John R. Hershey

Full-Capacity Unitary Recurrent Neural Networks
Full-Capacity Unitary Recurr ent Neural Networks Scott Wisdom 1 ∗ , Thomas Po wers 1 ∗ , John R. Hershey 2 , Jonathan Le Roux 2 , and Les Atlas 1 1 Department of Electrical Engineering, Univ ersity of W ashington {swisdom, tcpowers, atlas}@uw.edu 2 Mitsubishi Electric Research Laboratories (MERL) {hershey, leroux}@merl.com Abstract Recurrent neural networks are po werful models for processing sequential data, but they are generally plagued by vanishing and exploding gradient problems. Unitary recurrent neural networks (uRNNs), which use unitary recurrence matri- ces, hav e recently been proposed as a means to av oid these issues. Howe v er , in previous e xperiments, the recurrence matrices were restricted to be a product of parameterized unitary matrices, and an open question remains: when does such a parameterization fail to represent all unitary matrices, and how does this restricted representational capacity limit what can be learned? T o address this question, we propose full-capacity uRNNs that optimize their recurrence matrix over all unitary matrices, leading to significantly improv ed performance ov er uRNNs that use a restricted-capacity recurrence matrix. Our contribution consists of two main components. First, we provide a theoretical argument to determine if a unitary parameterization has restricted capacity . Using this argument, we show that a recently proposed unitary parameterization has restricted capacity for hidden state dimension greater than 7. Second, we sho w how a complete, full-capacity unitary recurrence matrix can be optimized over the dif ferentiable manifold of unitary matrices. The resulting multiplicative gradient step is very simple and does not require gradient clipping or learning rate adaptation. W e confirm the utility of our claims by empirically e v aluating our ne w full-capacity uRNNs on both synthetic and natural data, achie ving superior performance compared to both LSTMs and the original restricted-capacity uRNNs. 1 Introduction Deep feed-forward and recurrent neural networks ha ve been sho wn to be remarkably ef fectiv e in a wide variety of problems. A primary difficulty in training using gradient-based methods has been the so-called vanishing or exploding gr adient problem, in which the instability of the gradients over multiple layers can impede learning [ 1 , 2 ]. This problem is particularly keen for recurrent networks, since the repeated use of the recurrent weight matrix can magnify any instability . This problem has been addressed in the past by various means, including gradient clipping [ 3 ], using orthogonal matrices for initialization of the recurrence matrix [ 4 , 5 ], or by using pioneering architectures such as long short-term memory (LSTM) recurrent networks [ 6 ] or gated recurrent units [ 7 ]. Recently , sev eral innov ativ e architectures have been introduced to improve information flow in a network: residual networks, which directly pass information from previous layers up in a feed-forward network [ 8 ], and attention networks, which allow a recurrent network to access past acti vations [ 9 ]. The idea of using a unitary recurrent weight matrix w as introduced so that the gradients are inherently stable and do not vanish or explode [ 10 ]. The resulting unitary recurrent ∗ Equal contribution 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain. neural network (uRNN) is complex-v alued and uses a complex form of the rectified linear activ ation function. Howe v er , this idea was in v estigated using, as we show , a potentially restricted form of unitary matrices. The two main components of our contrib ution can be summarized as follows: 1) W e provide a theoretical argument to determine the smallest dimension N for which any parame- terization of the unitary recurrence matrix does not co ver the entire set of all unitary matrices. The argument relies on counting real-v alued parameters and using Sard’ s theorem to sho w that the smooth map from these parameters to the unitary manifold is not onto. Thus, we can sho w that a pre viously proposed parameterization [ 10 ] cannot represent all unitary matrices larger than 7 × 7 . Thus, such a parameterization results in what we refer to as a r estricted-capacity unitary recurrence matrix. 2) T o o vercome the limitations of restricted-capacity parameterizations, we propose a ne w method for stochastic gradient descent for training the unitary recurrence matrix, which constrains the gradient to lie on the differentiable manifold of unitary matrices. This approach allows us to directly optimize a complete, or full-capacity , unitary matrix. Neither restricted-capacity nor full-capacity unitary matrix optimization require gradient clipping. Furthermore, full-capacity optimization still achiev es good results without adaptation of the learning rate during training. T o test the limitations of a restricted-capacity representation and to confirm that our full-capacity uRNN does have practical implications, we test restricted-capacity and full-capacity uRNNs on both synthetic and natural data tasks. These tasks include synthetic system identification, long-term memorization, frame-to-frame prediction of speech spectra, and pixel-by-pixel classification of handwritten digits. Our proposed full-capacity uRNNs generally achiev e equiv alent or superior performance on synthetic and natural data compared to both LSTMs [ 6 ] and the original restricted- capacity uRNNs [10]. In the next section, we give an overvie w of unitary recurrent neural networks. Section 3 presents our first contribution: the theoretical argument to determine if any unitary parameterization has restricted-capacity . Section 4 describes our second contribution, where we sho w ho w to optimize a full-capacity unitary matrix. W e confirm our results with simulated and natural data in Section 5 and present our conclusions in Section 6. 2 Unitary recurr ent neural networks The uRNN proposed by Arjo vsky et al. [ 10 ] consists of the follo wing nonlinear dynamical system that has real- or complex-valued inputs x t of dimension M , complex-v alued hidden states h t of dimension N , and real- or complex-v alued outputs y t of dimension L : h t = σ b ( Wh t − 1 + Vx t ) y t = Uh t + c , (1) where y t = Re { Uh t + c } if the outputs y t are real-valued. The element-wise nonlinearity σ is [ σ b ( z )] i = ( ( | z i | + b i ) z i | z i | , if | z i | + b i > 0 , 0 , otherwise. (2) Note that this non-linearity consists in a soft-thresholding of the magnitude using the bias vector b . Hard-thresholding would set the output of σ to z i if | z i | + b i > 0 . The parameters of the uRNN are as follows: W ∈ U ( N ) , unitary hidden state transition matrix; V ∈ C N × M , input-to-hidden transformation; b ∈ R N , nonlinearity bias; U ∈ C L × N , hidden-to-output transformation; and c ∈ C L , output bias. Arjovsk y et al. [10] propose the following parameterization of the unitary matrix W : W u ( θ u ) = D 3 R 2 F − 1 D 2 PR 1 F D 1 , (3) where D are diagonal unitary matrices, R are Householder reflection matrices [ 11 ], F is a discrete Fourier transform (DFT) matrix, and P is a permutation matrix. The resulting matrix W u is unitary because all its component matrices are unitary . This decomposition is efficient because diagonal, reflection, and permutation matrices are O ( N ) to compute, and DFTs can be computed efficiently in O ( N log N ) time using the fast Fourier transform (FFT). The parameter vector θ u consists of 7 N 2 real-value d parameters: N parameters for each of the 3 diagonal matrices where D i,i = e j θ i and 2 N parameters for each of the 2 Householder reflection matrices, which are real and imaginary values of the complex reflection v ectors u i : R i = I − 2 u i u H i h u i , u i i . 3 Estimating the repr esentation capacity of structur ed unitary matrices In this section, we state and prov e a theorem that can be used to determine when any particular unitary parameterization does not ha ve capacity to represent all unitary matrices. As an application of this theorem, we sho w that the parameterization (3) does not ha ve the capacity to co ver all N × N unitary matrices for N > 7 . First, we establish an upper bound on the number of real-valued parameters required to represent any N × N unitary matrix. Then, we state and prove our theorem. Lemma 3.1 The set of all unitary matrices is a manifold of dimension N 2 . Proof: The set of all unitary matrices is the well-kno wn unitary Lie group U ( N ) [ 12 , §3.4]. A Lie group identifies group elements with points on a dif ferentiable manifold [ 12 , §2.2]. The dimension of the manifold is equal to the dimension of the Lie algebra u , which is a vector space that is the tangent space at the identity element [ 12 , §4.5]. For U ( N ) , the Lie algebra consists of all skew- Hermitian matrices A [ 12 , §5.4]. A ske w-Hermitian matrix is an y A ∈ C N × N such that A = − A H , where ( · ) H is the conjugate transpose. T o determine the dimension of U ( N ) , we can determine the dimension of u . Because of the ske w-Hermitian constraint, the diagonal elements of A are purely imaginary , which corresponds to N real-valued parameters. Also, since A i,j = − A ∗ j,i , the upper and lower triangular parts of A are parameterized by N ( N − 1) 2 complex numbers, which corresponds to an additional N 2 − N real parameters. Thus, U ( N ) is a manifold of dimension N 2 . Theorem 3.2 If a family of N × N unitary matrices is parameterized by P r eal-valued parameters for P < N 2 , then it cannot contain all N × N unitary matrices. Proof: W e consider a family of unitary matrices that is parameterized by P real-valued parameters through a smooth map g : P ( P ) → U ( N 2 ) from the space of parameters P ( P ) to the space of all unitary matrices U ( N 2 ) . The space P ( P ) of parameters is considered as a P -dimensional manifold, while the space U ( N 2 ) of all unitary matrices is an N 2 -dimensional manifold according to lemma 3.1. Then, if P < N 2 , Sard’ s theorem [ 13 ] implies that the image g ( P ) of g is of measure zero in U ( N 2 ) , and in particular g is not onto. Since g is not onto, there must exist a unitary matrix W ∈ U ( N 2 ) for which there is no corresponding input P ∈ P ( P ) such that W = g ( P ) . Thus, if P is such that P < N 2 , the manifold P ( P ) cannot represent all unitary matrices in U ( N 2 ) . W e no w apply Theorem 3.2 to the parameterization (3). Note that the parameterization (3) has P = 7 N real-valued parameters. If we solve for N in 7 N < N 2 , we get N > 7 . Thus, the parameterization (3) cannot represent all unitary matrices for dimension N > 7 . 4 Optimizing full-capacity unitary matrices on the Stiefel manif old In this section, we sho w how to get around the limitations of restricted-capacity parameterizations and directly optimize a full-capacity unitary matrix. W e consider the Stiefel manifold of all N × N complex-v alued matrices whose columns are N orthonormal vectors in C N [ 14 ]. Mathematically , the Stiefel manifold is defined as V N ( C N ) =  W ∈ C N × N : W H W = I N × N  . (4) For any W ∈ V N ( C N ) , any matrix Z in the tangent space T W V N ( C N ) of the Stiefel manifold satisfies Z H W − W H Z = 0 [ 14 ]. The Stiefel manifold becomes a Riemannian manifold when its tangent space is equipped with an inner product. T agare [ 14 ] suggests using the canonical inner product, giv en by h Z 1 , Z 2 i c = tr  Z H 1 ( I − 1 2 WW H ) Z 2  . (5) Under this canonical inner product on the tangent space, the gradient in the Stiefel manifold of the loss function f with respect to the matrix W is A W , where A = G H W − W H G is a ske w-Hermitian 3 matrix and G with G i,j = δ f δ W i,j is the usual gradient of the loss function f with respect to the matrix W [ 14 ]. Using these facts, T agare [ 14 ] suggests a descent curv e along the Stiefel manifold at training iteration k gi ven by the matrix product of the Cayle y transformation of A ( k ) with the current solution W ( k ) : Y ( k ) ( λ ) =  I + λ 2 A ( k )  − 1  I − λ 2 A ( k )  W ( k ) , (6) where λ is a learning rate and A ( k ) = G ( k ) H W ( k ) − W ( k ) H G ( k ) . Gradient descent proceeds by performing updates W ( k +1) = Y ( k ) ( λ ) . T agare [ 14 ] suggests an Armijo-W olfe search along the curve to adapt λ , but such a procedure w ould be expensiv e for neural network optimization since it requires multiple ev aluations of the forward model and gradients. W e found that simply using a fixed learning rate λ often works well. Also, RMSprop-style scaling of the gradient G ( k ) by a running av erage of the previous gradients’ norms [ 15 ] before applying the multiplicative step (6) can improv e con vergence. The only additional substantial computation required be yond the forward and backward passes of the network is the N × N matrix inv erse in (6). 5 Experiments All models are implemented in Theano [ 16 ], based on the implementation of restricted-capacity uRNNs by [ 10 ], av ailable from https://github.com/amarshah/complex_RNN . All code to replicate our results is av ailable from https://github.com/stwisdom/urnn . All models use RMSprop [ 15 ] for optimization, except that full-capacity uRNNs optimize their recurrence matrices with a fixed learning rate using the update step (6) and optional RMSprop-style gradient normalization. 5.1 Synthetic data First, we compare the performance of full-capacity uRNNs to restricted-capacity uRNNs and LSTMs on two tasks with synthetic data. The first task is synthetic system identification, where a uRNN must learn the dynamics of a target uRNN given only samples of the target uRNN’ s inputs and outputs. The second task is the copy memory problem, in which the network must recall a sequence of data after a long period of time. 5.1.1 System identification For the task of system identification, we consider the problem of learning the dynamics of a nonlinear dynamical system that has the form (1), giv en a dataset of inputs and outputs of the system. W e will draw a true system W sy s randomly from either a constrained set W u of restricted-capacity unitary matrices using the parameterization W u ( θ u ) in (3) or from a wider set W g of restricted-capacity unitary matrices that are guaranteed to lie outside W u . W e sample from W g by taking a matrix product of two unitary matrices drawn from W u . W e use a sequence length of T = 150 , and we set the input dimension M and output dimension L both equal to the hidden state dimension N . The input-to-hidden transformation V and output-to- hidden transformation U are both set to identity , the output bias c is set to 0 , the initial state is set to 0 , and the hidden bias b is drawn from a uniform distribution in the range [ − 0 . 11 , − 0 . 09] . The hidden bias has a mean of − 0 . 1 to ensure stability of the system outputs. Inputs are generated by sampling T -length i.i.d. sequences of zero-mean, diagonal and unit cov ariance circular complex- valued Gaussians of dimension N . The outputs are created by running the system (1) forward on the inputs. W e compare a restricted-capacity uRNN using the parameterization from (3) and a full-capacity uRNN using Stiefel manifold optimization with no gradient normalization as described in Section 4. W e choose hidden state dimensions N to test critical points predicted by our arguments in Section 3 of W u ( θ u ) in (3): N ∈ { 4 , 6 , 7 , 8 , 16 } . These dimensions are chosen to test below , at, and abov e the critical dimension of 7 . For all experiments, the number of training, validation, and test sequences are 20000 , 1000 , and 1000 , respectiv ely . Mean-squared error (MSE) is used as the loss function. The learning rate is 0 . 001 with a batch size of 50 for all experiments. Both models use the same matrix drawn from W u as initialization. T o isolate the ef fect of unitary recurrence matrix capacity , we only optimize W , setting 4 all other parameters to true oracle v alues. F or each method, we report the best test loss over 100 epochs and ov er 6 random initializations for the optimization. The results are sho wn in T able 1. “ W sy s init. ” refers to the initialization of the true system unitary matrix W sy s , which is sampled from either the restricted-capacity set W u or the wider set W g . T able 1: Results for system identification in terms of best normalized MSE. W u is the set of restricted-capacity unitary matrices from (3), and W g is a wider set of unitary matrices. W sy s init. Capacity N = 4 N = 6 N = 7 N = 8 N = 16 W u Restricted 4 . 81 e − 1 6 . 75 e − 3 3 . 53 e − 1 3 . 51 e − 1 7 . 30 e − 1 W u Full 1 . 28 e − 1 3 . 03 e − 1 2 . 16 e − 1 5 . 04 e − 2 1 . 28 e − 1 W g Restricted 3 . 21 e − 4 3 . 36 e − 1 3 . 36 e − 1 2 . 69 e − 1 7 . 60 e − 1 W g Full 8 . 72 e − 2 3 . 86 e − 1 2 . 62 e − 1 7 . 22 e − 2 1 . 00 e − 6 Notice that for N < 7 , the restricted-capacity uRNN achiev es comparable or better performance than the full-capacity uRNN. At N = 7 , the restricted-capacity and full-capacity uRNNs achie ve relati vely comparable performance, with the full-capacity uRNN achie ving slightly lower error . For N > 7 , the full-capacity uRNN always achie ves better performance versus the restricted-capacity uRNN. This result confirms our theoretical arguments that the restricted-capacity parameterization in (3) lacks the capacity to model all matrices in the unitary group for N > 7 and indicates the advantage of using a full-capacity unitary recurrence matrix. 5.1.2 Copy memory problem The experimental setup follows the copy memory problem from [ 10 ], which itself was based on the experiment from [ 6 ]. W e consider alternativ e hidden state dimensions and extend the sequence lengths to T = 1000 and T = 2000 , which are longer than the maximum length of T = 750 considered in previous literature. In this task, the data is a vector of length T + 20 and consists of elements from 10 cate gories. The vector begins with a sequence of 10 symbols sampled uniformly from categories 1 to 8. The next T − 1 elements of the vector are the ninth ’blank’ category , follo wed by an element from the tenth category , the ‘delimiter’. The remaining ten elements are ‘blank’. The task is to output T + 10 blank characters followed by the sequence from the be ginning of the vector . W e use average cross entrop y as the training loss function. The baseline solution outputs the blank category for T + 10 time steps and then guesses a random symbol uniformly from the first eight categories. This baseline has an expected a verage cross entropy of 10 log(8) T +20 . Figure 1: Results of the copy memory problem with sequence lengths of 1000 (left) and 2000 (right). The full-capacity uRNN conv erges quickly to a perfect solution, while the LSTM and restricted- capacity uRNN with approximately the same number of parameters are unable to improv e past the baseline naiv e solution. The full-capacity uRNN uses a hidden state size of N = 128 with no gradient normalization. T o match the number of parameters ( ≈ 22k ), we use N = 470 for the restricted-capacity uRNN, and N = 68 for the LSTM. The training set size is 100000 and the test set size is 10000. The results 5 of the T = 1000 experiment can be found on the left half of Figure 1. The full-capacity uRNN con verges to a solution with zero a verage cross entropy after about 2000 training iterations, whereas the restricted-capacity uRNN settles to the baseline solution of 0.020. The results of the T = 2000 experiment can be found on the right half of Figure 1. The full-capacity uRNN hovers around the baseline solution for about 5000 training iterations, after which it drops do wn to zero av erage cross entropy . The restricted-capacity again settles do wn to the baseline solution of 0.010. These results demonstrate that the full-capacity uRNN is very ef fective for problems requiring v ery long memory . 5.2 Speech data W e now apply restricted-capacity and full-capacity uRNNs to real-world speech data and compare their performance to LSTMs. The main task we consider is predicting the log-magnitude of future frames of a short-time Fourier transform (STFT). The STFT is a commonly used feature domain for speech enhancement, and is defined as the Fourier transform of short windo wed frames of the time series. In the STFT domain, a real-valued audio signal is represented as a complex-v alued F × T matrix composed of T frames that are each composed of F = N win / 2 + 1 frequency bins, where N win is the duration of the time-domain frame. Most speech processing algorithms use the log-magnitude of the complex STFT values and reconstruct the processed audio signal using the phase of the original observations. The frame prediction task is as follo ws: gi ven all the log-magnitudes of STFT frames up to time t , predict the log-magnitude of the STFT frame at time t + 1 .W e use the TIMIT dataset [ 17 ]. According to common practice [ 18 ], we use a training set with 3690 utterances from 462 speakers, a v alidation set of 400 utterances, an e valuation set of 192 utterances. T raining, validation, and e valuation sets hav e distinct speakers. Results are reported on the ev aluation set using the network parameters that perform best on the v alidation set in terms of the loss function over three training trials. All TIMIT audio is resampled to 8 kHz. The STFT uses a Hann analysis window of 256 samples ( 32 milliseconds) and a window hop of 128 samples ( 16 milliseconds). The LSTM requires gradient clipping during optimization, while the restricted-capacity and full- capacity uRNNs do not. The hidden state dimensions N of the LSTM are chosen to match the number of parameters of the full-capacity uRNN. For the restricted-capacity uRNN, we run models that match either N or number of parameters. For the LSTM and restricted-capacity uRNNs, we use RMSprop [ 15 ] with a learning rate of 0 . 001 , momentum 0 . 9 , and a veraging parameter 0 . 1 . For the full-capacity uRNN, we also use RMSprop to optimize all network parameters, except for the recurrence matrix, for which we use stochastic gradient descent along the Stiefel manifold using the update (6) with a fixed learning rate of 0 . 001 and no gradient normalization. T able 2: Log-magnitude STFT prediction results on speech data, ev aluated using objectiv e and perceptual metrics (see text for description). Model N # parameters V alid. MSE Eval. MSE SegSNR (dB) STOI PESQ LSTM 84 ≈ 83k 18.02 18.32 1.95 0.77 1.99 Restricted-capacity uRNN 128 ≈ 67k 15.03 15.78 3.30 0.83 2.36 Restricted-capacity uRNN 158 ≈ 83k 15.06 14.87 3.32 0.83 2.33 Full-capacity uRNN 128 ≈ 83k 14.78 15.24 3.57 0.84 2.40 LSTM 120 ≈ 135k 16.59 16.98 2.32 0.79 2.14 Restricted-capacity uRNN 192 ≈ 101k 15.20 15.17 3.31 0.83 2.35 Restricted-capacity uRNN 256 ≈ 135k 15.27 15.63 3.31 0.83 2.36 Full-capacity uRNN 192 ≈ 135k 14.56 14.66 3.76 0.84 2.42 LSTM 158 ≈ 200k 15.49 15.80 2.92 0.81 2.24 Restricted-capacity uRNN 378 ≈ 200k 15.78 16.14 3.16 0.83 2.35 Full-capacity uRNN 256 ≈ 200k 14.41 14.45 3.75 0.84 2.38 Results are shown in T able 2, and Figure 2 shows e xample predictions of the three types of networks. Results in T able 2 are gi ven in terms of the mean-squared error (MSE) loss function and sev eral metrics computed on the time-domain signals, which are reconstructed from the predicted log-magnitude 6 Figure 2: Ground truth and one-frame-ahead predictions of a spectrogram for an example utterance. For each model, hidden state dimension N is chosen for the best v alidation MSE. Notice that the full-capacity uRNN achiev es the best detail in its predictions. and the original phase of the STFT . These time-domain metrics are se gmental signal-to-noise ratio (SegSNR), short-time objecti ve intelligibility (STOI), and perceptual e valuation of speech quality (PESQ). SegSNR, computed using [ 19 ], uses a voice activity detector to av oid measuring SNR in silent frames. STOI is designed to correlate well with human intelligibility of speech, and tak es on values between 0 and 1, with a higher score indicating higher intelligibility [ 20 ]. PESQ is the ITU-T standard for telephone v oice quality testing [ 21 , 22 ], and is a popular perceptual quality metric for speech enhancement [23]. PESQ ranges from 1 (bad quality) to 4.5 (no distortion). Note that full-capacity uRNNs generally perform better than restricted-capacity uRNNs with the same number of parameters, and both types of uRNN significantly outperform LSTMs. 5.3 Pixel-by-pixel MNIST As another challenging long-term memory task with natural data, we test the performance of LSTMs and uRNNs on pixel-by-pix el MNIST and permuted pixel-by-pix el MNIST , first proposed by [ 5 ] and used by [ 10 ] to test restricted-capacity uRNNs. For permuted pixel-by-pixel MNIST , the pix els are shuf fled, thereby creating some non-local dependencies between pixels in an image. Since the MNIST images are 28 × 28 pixels, resulting pix el-by-pixel sequences are T = 784 elements long. W e use 5000 of the 60000 training examples as a validation set to perform early stopping with a patience of 5. The loss function is cross-entropy . W eights with the best validation loss are used to process the ev aluation set. The full-capacity uRNN uses RMSprop-style gradient normalization. T able 3: Results for unpermuted and permuted pix el-by-pixel MNIST . Classification accuracies are reported for trained model weights that achiev e the best validation loss. Model N # parameters V alidation accurary Evaluation accuracy Unpermuted LSTM 128 ≈ 68k 98.1 97.8 LSTM 256 ≈ 270k 98.5 98.2 Restricted-capacity uRNN 512 ≈ 16k 97.9 97.5 Full-capacity uRNN 116 ≈ 16k 92.7 92.8 Full-capacity uRNN 512 ≈ 270k 97.5 96.9 Permuted LSTM 128 ≈ 68k 91.7 91.3 LSTM 256 ≈ 270k 92.1 91.7 Restricted-capacity uRNN 512 ≈ 16k 94.2 93.3 Full-capacity uRNN 116 ≈ 16k 92.2 92.1 Full-capacity uRNN 512 ≈ 270k 94.7 94.1 7 Figure 3: Learning curves for unpermuted pixel-by-pix el MNIST (top panel) and permuted pixel-by- pixel MNIST (bottom panel). Learning curves are sho wn in Figure 3, and a summary of classification accuracies is shown in T able 3. For the unpermuted task, the LSTM with N = 256 achiev es the best e valuation accuracy of 98 . 2% . For the permuted task, the full-capacity uRNN with N = 512 achiev es the best ev aluation accuracy of 94 . 1% , which is state-of-the-art on this task. Both uRNNs outperform LSTMs on the permuted case, achie ving their best performance after fe wer traing epochs and using an equal or lesser number of trainable parameters. This performance difference suggests that LSTMs are only able to model local dependencies, while uRNNs ha ve superior long-term memory capabilities. Despite not representing all unitary matrices, the restricted-capacity uRNN with N = 512 still achie ves impressiv e test accuracy of 93 . 3% with only 1 / 16 of the trainable parameters, outperforming the full-capacity uRNN with N = 116 that matches number of parameters. This result suggests that further exploration into the potential trade-off between hidden state dimension N and capacity of unitary parameterizations is necessary . 6 Conclusion Unitary recurrent matrices prove to be an ef fective means of addressing the v anishing and exploding gradient problems. W e provided a theoretical argument to quantify the capacity of constrained unitary matrices. W e also described a method for directly optimizing a full-capacity unitary matrix by constraining the gradient to lie in the differentiable manifold of unitary matrices. The ef fect of restricting the capacity of the unitary weight matrix was tested on system identification and memory tasks, in which full-capacity unitary recurrent neural networks (uRNNs) outperformed restricted- capacity uRNNs from [ 10 ] as well as LSTMs. Full-capacity uRNNs also outperformed restricted- capacity uRNNs on log-magnitude STFT prediction of natural speech signals and classification of permuted pixel-by-pixel images of handwritten digits, and both types of uRNN significantly outperformed LSTMs. In future work, we plan to explore more general forms of restricted-capacity unitary matrices, including constructions based on products of elementary unitary matrices such as Householder operators or Giv ens operators. Acknowledgments: W e thank an anonymous re viewer for suggesting impro vements to our proof in Section 3 and V amsi Potluru for helpful discussions. Scott W isdom and Thomas Powers were funded by U.S. ONR contract number N00014-12-G-0078, delivery orders 13 and 24. Les Atlas was funded by U.S. AR O grant W911NF-15-1-0450. 8 References [1] Y . Bengio, P . Simard, and P . Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE T ransactions on Neural Networks , 5(2):157–166, 1994. [2] S. Hochreiter , Y . Bengio, P . Frasconi, and J. Schmidhuber . Gradient flow in recurrent nets: the difficulty of learning long-term dependencies. In S. C. Kremer and J. F . K olen, eds, A field guide to dynamical r ecurr ent neural networks . IEEE Press, 2001. [3] R. Pascanu, T . Mikolov , and Y . Bengio. On the difficulty of training Recurrent Neural Networks. arXiv:1211.5063 , Nov . 2012. [4] A. M. Saxe, J. L. McClelland, and S. Ganguli. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. , Dec. 2013. [5] Q. V . Le, N. Jaitly , and G. E. Hinton. A simple way to initialize recurrent networks of rectified linear units. arXiv:1504.00941 , Apr . 2015. [6] S. Hochreiter and J. Schmidhuber . Long short-term memory . Neural computation , 9(8):1735–1780, 1997. [7] K. Cho, B. van Merriënboer , D. Bahdanau, and Y . Bengio. On the properties of neural machine translation: Encoder-decoder approaches. , 2014. [8] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. , Dec. 2015. [9] V . Mnih, N. Heess, A. Grav es, and K. Kavukcuoglu. Recurrent models of visual attention. In Advances in Neural Information Pr ocessing Systems (NIPS) , pp. 2204–2212, 2014. [10] M. Arjovsky , A. Shah, and Y . Bengio. Unitary Evolution Recurrent Neural Networks. In International Confer ence on Machine Learning (ICML) , Jun. 2016. [11] A. S. Householder . Unitary triangularization of a nonsymmetric matrix. Journal of the A CM , 5(4):339–342, 1958. [12] R. Gilmore. Lie gr oups, physics, and geometry: an intr oduction for physicists, engineers and chemists . Cambridge Univ ersity Press, 2008. [13] A. Sard. The measure of the critical values of dif ferentiable maps. Bulletin of the American Mathematical Society , 48(12):883–890, 1942. [14] H. D. T agare. Notes on optimization on Stiefel manifolds. T echnical report, Y ale Univ ersity , 2011. [15] T . Tieleman and G. Hinton. Lecture 6.5—RmsProp: Divide the gradient by a running average of its recent magnitude, 2012. COURSERA: Neural Networks for Machine Learning. [16] Theano Dev elopment T eam. Theano: A Python framework for fast computation of mathematical e xpres- sions. arXiv: 1605.02688 , May 2016. [17] J. S. Garofolo, L. F . Lamel, W . M. Fisher , J. G. Fiscus, and D. S. Pallett. D ARP A TIMIT acoustic-phonetic continous speech corpus. T echnical Report NISTIR 4930, National Institute of Standards and T echnology , 1993. [18] A. K. Halberstadt. Hetero geneous acoustic measur ements and multiple classifiers for speech r ecognition . PhD thesis, Massachusetts Institute of T echnology , 1998. [19] M. Brookes. V OICEBOX: Speech processing toolbox for MA TLAB, 2002. [Online]. A vailable: http://www .ee.ic.ac.uk/hp/staff/dmb/v oicebox/voicebox.html. [20] C. T aal, R. Hendriks, R. Heusdens, and J. Jensen. An algorithm for intelligibility prediction of time- frequency weighted noisy speech. IEEE T rans. on Audio, Speec h, and Language Pr ocessing , 19(7):2125– 2136, Sep. 2011. [21] A. Rix, J. Beerends, M. Hollier , and A. Hekstra. Perceptual ev aluation of speech quality (PESQ)-a ne w method for speech quality assessment of telephone networks and codecs. In Pr oc. ICASSP , vol. 2, pp. 749–752, 2001. [22] ITU-T P .862. Perceptual ev aluation of speech quality (PESQ): An objective method for end-to-end speech quality assessment of narrow-band telephone netw orks and speech codecs, 2000. [23] P . C. Loizou. Speech Enhancement: Theory and Practice . CRC Press, Boca Raton, FL, Jun. 2007. 9

Original Paper

Loading high-quality paper...

Comments & Academic Discussion

Loading comments...

Leave a Comment