Class-conditional embeddings for music source separation
Isolating individual instruments in a musical mixture has a myriad of potential applications, and seems imminently achievable given the levels of performance reached by recent deep learning methods. While most musical source separation techniques lea…
Authors: Prem Seetharaman, Gordon Wichern, Shrikant Venkataramani
CLASS-CONDITIONAL EMBEDDINGS FOR MUSIC SOURCE SEP ARA TION Pr em Seetharaman 1 , 2 , Gordon W ic hern 1 , Shrikant V enkatar amani 1 , 3 , Jonathan Le Roux 1 1 Mitsubishi Electric Research Laboratories (MERL), Cambridge, MA, USA 2 Northwestern Uni versity , Evanston, IL, USA 3 Uni versity of Illinois at Urbana-Champaign, Champaign, IL, USA ABSTRA CT Isolating individual instruments in a musical mixture has a myriad of potential applications, and seems imminently achiev able giv en the lev els of performance reached by recent deep learning meth- ods. While most musical source separation techniques learn an in- dependent model for each instrument, we propose using a common embedding space for the time-frequency bins of all instruments in a mixture inspired by deep clustering and deep attractor networks. Additionally , an auxiliary network is used to generate parameters of a Gaussian mixture model (GMM) where the posterior distrib ution ov er GMM components in the embedding space can be used to cre- ate a mask that separates indi vidual sources from a mixture. In addi- tion to outperforming a mask-inference baseline on the MUSDB-18 dataset, our embedding space is easily interpretable and can be used for query-based separation. Index T erms — source separation, deep clustering, music, classifi- cation, neural networks 1. INTR ODUCTION Audio source separation is the act of isolating sound-producing sources in an auditory scene. Examples include separating singing voice from accompanying music, the voice of a single speaker at a crowded party , or the sound of a car backfiring in a loud urban soundscape. Recent deep learning techniques hav e rapidly adv anced the performance of such source separation algorithms leading to state of the art performance in the separation of music mixtures [1], separation of speech from non-stationary background noise [2], and separation of the voices from simultaneous overlapping speak- ers [3], often using only a single audio channel as input, i.e., no spatial information. In this work we are concerned with separation networks that take as input a time-frequency (T -F) representation of a signal (e.g., magni- tude spectrogram), and either predict the separated source v alue in each T -F bin directly or via a T -F mask that when multiplied with the input recovers the separated signal. An inv erse transform is then used to obtain the separated audio. One approach for training such algorithms uses some type of signal reconstruction error , such as the mean square error between magnitude spectra [2, 4]. An alternative approach referred to as deep clustering [3, 5, 6] uses affinity-based training by estimating a high-dimensional embedding for each T -F bin, training the network with a loss function such that the embed- dings for T -F bins dominated by the same source should be close to each other and those for bins dominated by different sources should This work was performed while P . Seetharaman and S. V enkataramani were interns at MERL. BLSTM stack Embedding space 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 μ 0 , Σ 0 , π 0 μ 1 , Σ 1 , π 1 μ 3 , Σ 3 , π 3 Auxiliary class conditional network μ 2 , Σ 2 , π 2 V ocals Drums Bass Other ℒ L 1 ℒ DC ℒ L 1 ℒ L 1 ℒ L 1 Fig. 1 . Class-conditional embedding network architecture. An aux- iliary network generates the parameters of Gaussians in the embed- ding space, taking a one-hot vector indicating the class as input. The masks are generated using the posteriors of the Gaussian mixture model across all the sources. The network is trained using a deep clustering loss ( L DC ) on the embedding space and an L 1 loss ( L L 1 ) on the masked spectrograms. be far apart. This affinity-based training is especially valuable in tasks such as speech separation, as it a voids the permutation problem during network training where there is no straightforward mapping between the order of targets and outputs. Deep clustering for music source separation was previously inv es- tigated with Chimera netw orks for singing voice separation in [7]. Chimera networks [7, 6] have multiple parallel output heads which are trained simultaneously on different tasks. Specifically , in the case of singing voice separation, one output head is used to directly ap- proximate the soft mask for extracting vocals, while the other head outputs an embedding space that optimizes the deep clustering loss. When both heads are trained together , results are better than using any single head alone. Another approach for combining deep clus- tering and mask-based techniques was presented in [5] where a deep clustering network, unfolded k-means layers, and a second stage en- hancement network are trained end-to-end. The deep attractor net- work [8] computes an embedding for each T -F bin similar to deep clustering, but creates a mask based on the distance of each T -F bin to an attractor point for each source. The attractors can either be estimated via k-means clustering, or learned as fix ed points during training. In this work, we consider deep attractor-lik e networks for separating multiple instruments in music mixtures. While embedding networks hav e typically been used in speech separation where all sources in a mixture belong to the same class (human speak ers), we extend the formulation to situations where sources in a mixture correspond to distinct classes (e.g., musical instruments). Specifically , our class- conditional embeddings work as follows: first, we propose using an auxiliary network to estimate a Gaussian distribution (mean v ector and cov ariance matrix) in an embedding space for each instrument class we are trying to separate. Then, another network computes an embedding for each T -F bin in a mixture, akin to deep clustering. Finally , a mask is generated based on the posterior distribution over classes for each T -F bin. The network can be trained using a signal reconstruction objectiv e, or in a multi-task (Chimera-like) fashion with an affinity-based deep clustering loss used as a re gularizer . Deep clustering and deep attractor networks typically focus on speaker -independent speech separation where the mapping of input speaker to output index is treated as a nuisance parameter handled via permutation free training [3, 5, 9]. Sev eral recent works on speaker -conditioned separation [10, 11] allow separation of a tar- geted speaker from a mixture in a manner similar to how we extract specific instruments. Learning an embedding space for speaker separation that could also be used for classification was explored in recent work [12]. Howe ver , their work did this by introducing a classification-based loss function. Here, the conditioning is intro- duced as input into the network rather than output from the network. Regarding specific musical instrument extraction from mixtures, a majority of methods [13, 14, 15, 1] use an independent deep net- work model for each instrument, and then combine these instrument specific network outputs in post-processing using a technique such as the multi-channel W iener filter [14, 15]. While the efficacy of independent instrument modeling for musical source separation was confirmed by the results of a recent challenge [16], the requirements both in terms of computational resources and training data can be large, and scaling up the number of possible instruments can be prohibitiv e. Recently , the work in [17] demonstrated that a common embedding space for musical instrument separation using various deep attrac- tor networks could achiev e competiti ve performance. Our system is similar to the anchored and/or expectation-maximization deep at- tractor networks in [17], but we use an auxiliary netw ork to estimate the mean and co variance parameters for each instrument. W e also explore what type of covariance model is most effecti ve for musi- cal source separation (tied vs. untied across classes, diagonal vs. spherical). Furthermore, we discuss a simple modification of our pre-trained embedding networks for query-by-example separation [18, 19, 20], where given an isolated example of a sound we w ant to separate, we can extract the portion of a mixture most like the query without supervision. 2. EMBEDDING NETWORKS Let X ∈ C F × T be the complex spectrogram of the mixture of C sources S c ∈ C F × T for c = 1 , . . . , C . An embedding network computes V = f ( ˜ X ) (1) where ˜ X is the input feature representation (we use the log- magnitude spectrogram in this work), and V ∈ R F T × K contains a K -dimensional embedding for each T -F bin in the spectrogram. The function f is typically a deep neural netw ork composed of bidi- rectional long short-term memory (BLSTM) layers, followed by a dense layer . W e then create a mask for each source m c ∈ R F T × 1 , with C X c =1 m c,j = 1 , j = 1 , . . . , F T (2) from V . Deep clustering [3, 5] builds binary masks via k-means clustering on V (soft masks can also be obtained via soft k-means), and is trained by minimizing the dif ference between the true and estimated affinity matrices, L DC ( V , Y ) = k VV T − YY T k 2 F (3) where Y ∈ R T F × C indicates which of the C sources dominates each T -F bin. Deep attractor networks [8] use the distance between the embeddings and fixed attractor points in the embedding space to compute soft masks, and are typically trained with a signal recon- struction loss function, such as the L 1 loss between the estimated and ground truth magnitude spectrograms L L 1 = C X c =1 | m c x − s c | . (4) where x , s c ∈ R F T × 1 are the flattened spectrogram of the mixture, and ground truth source, respecti vely . W e can obtain the separated time domain signal from the estimated magnitude ˆ s c = m c x after an in verse STFT using the mixture phase. Chimera networks [7, 6] combine signal reconstruction and deep clustering losses, using two heads stemming from the same under - lying network (stacked BLSTM layers). In this work, we also com- bine the loss functions from (3) and (4) (with equal weighting), but the gradients from both propagate into the same embedding space, rather than separate heads. 3. CONDITIONING EMBEDDINGS ON CLASS When we are interested in separating sources that belong to distinctly different groups, i.e., classes, each source c has an associated class label z c , and we assume here that each mixture contains at most one isolated source per class label. Estimating the mask m c,j in (2) for source (class) c and T -F bin j is then equiv alent to estimating the pos- terior o ver classes p ( z c | v j ) given the corresponding K -dimensional network embedding. F or simplicity we use a Gaussian model of the embedding space and obtain the mask from m c,j = p ( z c | v j ) = π c N ( v j | µ µ µ c , Σ c ) P C i =1 π i N ( v j | µ µ µ i , Σ i ) . (5) The Gaussian parameters ( µ µ µ c , Σ c ) and class prior π c for each class are learned end-to-end along with the embedding network. The gen- eration of the parameters of each Gaussian from the auxiliary class- conditional network is the maximization step in the expectation- maximization (EM) algorithm (trained through stochastic gradient descent), and the generation of the mask is the expectation step. Rather than unfolding a clustering algorithm as in [5], we instead can learn the parameters of the clustering algorithm ef ficiently via gradient descent. Further, the soft mask is generated directly from the posteriors of the Gaussians, rather than through a second-stage enhancement network as in [5]. A diagram of our system can be seen in Fig. 1. W e also draw a connection between the class-conditional masks of (5) and the adaptiv e pooling layers for sound ev ent detection in [21], which are also conditioned on class label. In [21], an activ ation function that is a variant of softmax with learnable parameter α is introduced. If α is very high, the function reduces to a max func- tion, heavily emphasizing the most likely class. If it is low , energy Fig. 2 . V isualization of the embedding space learned for music source separation with the tied spherical cov ariance model. is spread more e venly across classes approximating an av erage. Our work uses a similar idea for source separation. A softmax nonlin- earity is comparable to the posterior probability computation used in the e xpectation step of the EM algorithm in our Gaussian mixture model (GMM). For a GMM with tied spherical cov ariance, α is the in verse of the variance. A similar formulation of softmax was also used in [5], where k-means was unfolded on an embedding space. In that w ork α was set manually to a high value for good results. In our work, we effectiv ely learn the optimal α (the inv erse of the covari- ance matrix) for signal reconstruction rather than setting it manually , but still conditioning it on class as in [21] for source separation. 4. EXPERIMENTS Our experiment is designed to in vestigate whether our proposed class-conditional model outperforms a baseline mask inference model. W e also explore which covariance type is most suitable for music source separation. W e do this by ev aluating the SDR of sepa- rated estimates of vocals, drums, bass, and other in the MUSDB [22] corpus using the museval package 1 . Finally , we show the potential of our system to perform querying tasks with isolated sources. 4.1. Dataset and training procedur e W e e xtend Scaper [23], a library for soundscape mixing designed for sound e vent detection, to create large synthetic datasets for source separation. W e apply our variant of Scaper to the MUSDB train- ing data, which consists of 100 songs with vocals, drums, bass, and other stems to create 20000 training mixtures and 2000 validation mixtures, all of length 3 . 2 seconds at a sampling rate of 48 kHz. Of the 100 songs, we use 86 for training and 14 for validation. The re- maining 50 songs in the MUSDB testing set are used for testing. The 1 https://github .com/sigsep/sigsep-mus-ev al training and validation set mixtures are musically incoherent (ran- domly created using stems from different songs) and each contains a random 3 . 2 second excerpt from a stem audio file in MUSDB (vo- cals, drums, bass, and other). All four sources are present in ev ery training and validation mixture. T ime domain stereo audio is summed to mono and transformed to a single-channel log-magnitude spectrogram with a window size of 2048 samples ( ≈ 43 ms) and a hop size of 512 samples. Our net- work consists of a stack of 4 BLSTMs layers with 300 units in each direction for a total of 600 . Before the BLSTM stack, we project the log-magnitude spectrogram to a mel-spectrogram with 300 mel bins. The mel spectrogram frames are fed to the BLSTM stack which projects ev ery time-mel bin to an embedding with 15 dimensions. The auxiliary class-conditional network takes as input a one-hot vec- tor of size 4 , one for each musical source class in our dataset. It maps the one-hot v ector to the parameters of a Gaussian in the embedding space. For an embedding space of size K , and diagonal covariance matrix, the one-hot vector is mapped to a vector of size 2 K + 1 : K for the mean, K for the variance, and 1 for the prior . After the pa- rameters of all Gaussians are generated, we compute the mask from the posteriors across the GMM using Eq. (5). The resultant mask is then put through an inv erse mel transform to project it back to the linear frequency domain, clamped between 0 and 1 , and applied to the mixture spectrogram. The system is trained end to end with L 1 loss and the embedding space is regularized using the deep cluster- ing loss function. T o compute the deep clustering loss, we need the affinity matrix for the mel-spectrogram space. This is computed by projecting the ideal binary masks for each source into mel space and clamping between 0 and 1 . The deep clustering loss is only applied on bins that hav e a log magnitude louder than − 40 db, following [3]. W e evaluate the performance of multiple variations of class-conditional embedding networks on the MUSDB18 [22] dataset using source- to-distortion ratio (SDR) 2 . At test time, we apply our network to both stereo channels independently and mask the two channels of complex stereo spectrogram. W e explore sev eral v ariants of our system, specifically focusing on the possible cov ariance shapes of the learned Gaussians. W e compare these models to a baseline model that is simply a mask inference network (the same BLSTM stack) with 4 F outputs (one mask per class) followed by a sigmoid activ ation. All networks start from the same initialization and are trained on the same data. 4.2. Results T able 1 shows SDR results for the baseline model and the four co- variance model variants. W e find that all four of our models that use an embedding space improv e significantly on the baseline for vo- cals and other sources. The best performing model is a GMM with tied spherical cov ariance, which reduces to soft k-means, as used in [5]. The difference here is that the v alue for the covariance is learned rather than set manually . The covariance learned was 0 . 16 , or an α value of 6 , close to the α value of 5 found in [5]. The em- bedding space for this model on a sample mixture is visualized in Fig. 2 using Principal Component Analysis. W e observe that there exist “bridges” between some of the sources. For example, other and vocals share many time-frequency points, possibly due to their source similarity . Both sources contain harmonic sounds and some- times leading melodic instruments. Ho wever , unlik e other embed- 2 https://github .com/sigsep/sigsep-mus-ev al Fig. 3 . An example of an embedding dimension for a GMM with diagonal covariance. The embedding selected here is the one with the lowest v ariance ( 0 . 04 ) for the vocals source. T able 1 . SDR comparing mask inference baseline (BLSTM) with deep clustering (DC) plus GMM models with different covariance types (diagonal, tied diagonal, spherical, and tied spherical). Approach V ocals Drums Bass Other BLSTM 3 . 82 4 . 14 2 . 48 2 . 35 DC/GMM - diag. (untied) 4 . 20 4 . 26 2 . 58 2 . 55 DC/GMM - diag. (tied) 4 . 04 3 . 96 2 . 48 2 . 47 DC/GMM - sphr . (untied) 4 . 21 4 . 19 2 . 29 2 . 58 DC/GMM - sphr . (tied) 4 . 49 4 . 23 2 . 73 2 . 51 ding spaces (e.g., word2v ec) where things that are similar are near each other in the embedding space, we instead ha ve learned a sep- aration space, where sources that are similar (but different) seem to be placed far from each other in the embedding space. W e hypoth- esize that this is to optimize the separation objecti ve. In [8], it is observed that attractors for speaker separation come in two pairs, across from each other . Our work suggests that the two pairs may correspond to similar sources (e.g., separating female speakers from one another and separating male speakers from one another). V erify- ing this and understanding embedding spaces learned by embedding networks will be the subject of future work. W e hypothesize that the reason the simplest covariance model (tied spherical) performs best in T able 1 is that for the diagonal case, the variances collapse in all but a few embedding dimensions. Em- bedding dimensions with the lowest variance contribute most to the ov erall mask. As a result, they essentially become the mask by them- selves, reducing the network more to mask inference rather than an embedding space. An example of this can be seen in Fig. 3, where the embedding dimension has essentially reduced to a mask for the vocals source. W ith a spherical covariance model, each embedding dimension must be treated equally , and the embedding space cannot collapse to mask inference. A possible reason tied spherical per- forms better than untied spherical, may be that the network becomes ov erly confident (low variance) for certain classes. W ith a tied spher - ical cov ariance structure, all embedding dimensions and instrument classes are equally weighted, forcing the network to use more of the embedding space, perhaps leading to better performance. 4.3. Querying with isolated sources T o query a mixture with isolated sources, we propose a simple ap- proach that leverages an already trained class-conditional embedding network. W e take the query audio and pass it through the network to produce an embedding space. Then, we fit a Gaussian with a sin- gle component to the resultant embedding space. Next, we take a Query (snare) Mixture Query mask applied to mixture Mask produced by query Gaussian Fig. 4 . An example of query by isolated source using our trained tied-spherical model. Upper left: mixture spectrogram, upper right: query (snare drum) spectrogram, bottom right: query mask, bottom left: masked mixture using query mask. mixture that may contain audio of the same type as the query , but not the exact instance of the query , and pass that through the same network. This produces an embedding for the mixture. T o extract similar content to the query from the mixture, we take the Gaussian that was fit to the query embeddings and run the expectation step of EM by calculating the likelihood of the mixture’ s embeddings under the query’ s Gaussian. Because there is only one component in this mixture model, calculating posteriors gives a mask of all ones. T o alleviate this, we use the likelihood under the query Gaussian as the mask on the mixture and normalize it to [0 , 1] by dividing each lik e- lihood by the maximum observed likelihood v alue in the mixture. An example of query by isolated source can be seen in Fig. 4. W e use a recording of solo snare drum as our query . The snare drum in the query is from an unrelated recording found on Y ouT ube. The mix- ture recording is of a song with simultaneous vocals, drums, bass, and guitar ( Heart of Gold - Neil Y oung). The Gaussian is fit to the snare drum embeddings and transferred to the mixture embeddings. The mask produced is similar to the query as is the extracted part of the mixture. This inv ariance of embedding location was a result of conditioning the embeddings on the class. 5. CONCLUSION W e have presented a method for conditioning on class an embed- ding space for source separation. W e ha ve extended the formula- tion of deep attractor networks and other embedding networks to accommodate Gaussian mixture models with different cov ariances. W e test our method on musical mixtures and found that it outper- forms a mask inference baseline. W e find that the embeddings found by the network are interpretable to an extent and hypothesize that embeddings are learned such that source classes that hav e similar characteristics are kept far from each other in order to optimize the separation objectiv e. Our model can be easily adapted to a querying task using an isolated source. In future work, we hope to in vestigate the dynamics of embedding spaces for source separation, apply our approach to more general audio classes, and explore the querying task further . 6. REFERENCES [1] N. T akahashi, N. Goswami, and Y . Mitsufuji, “MMDenseL- STM: An efficient combination of con volutional and recurrent neural networks for audio source separation, ” in Pr oc. Interna- tional W orkshop on Acoustic Signal Enhancement (IW AENC) , 2018. [2] H. Erdogan, J. R. Hershey , S. W atanabe, and J. Le Roux, “Phase-sensitiv e and recognition-boosted speech separation using deep recurrent neural networks, ” in Pr oc. IEEE Interna- tional Conference on Acoustics, Speech and Signal Pr ocessing (ICASSP) , 2015. [3] J. R. Hershey , Z. Chen, and J. Le Roux, “Deep clustering: Dis- criminativ e embeddings for segmentation and separation, ” in Pr oc. IEEE International Conference on Acoustics, Speech and Signal Pr ocessing (ICASSP) , 2016. [4] D. W ang and J. Chen, “Supervised speech separation based on deep learning: An overvie w , ” in arXiv preprint arXiv:1708.07524 , 2017. [5] Y . Isik, J. Le Roux, Z. Chen, S. W atanabe, and J. R. Hershey , “Single-channel multi-speaker separation using deep cluster- ing, ” arXiv preprint , 2016. [6] Z.-Q. W ang, J. Le Roux, and J. R. Hershe y , “ Alternative ob- jectiv e functions for deep clustering, ” in Pr oc. IEEE Interna- tional Conference on Acoustics, Speech and Signal Pr ocessing (ICASSP) , 2018. [7] Y . Luo, Z. Chen, J. R. Hershey , J. Le Roux, and N. Mesgarani, “Deep clustering and con ventional networks for music sepa- ration: Stronger together, ” in Acoustics, Speech and Signal Pr ocessing (ICASSP), 2017 IEEE International Conference on . IEEE, 2017. [8] Y . Luo, Z. Chen, and N. Mesgarani, “Speaker-independent speech separation with deep attractor network, ” IEEE/ACM T ransactions on Audio, Speech, and Language Processing , vol. 26, no. 4, 2018. [9] M. Kolbæk, D. Y u, Z.-H. T an, and J. Jensen, “Multi-talker speech separation with utterance-le vel permutation in variant training of deep recurrent neural networks, ” IEEE/ACM T rans- actions on Audio, Speech and Language Pr ocessing , 2017. [10] M. Delcroix, K. Zmolikov a, K. Kinoshita, A. Ogawa, and T . Nakatani, “Single channel target speaker extraction and recognition with speaker beam, ” in 2018 IEEE Interna- tional Conference on Acoustics, Speech and Signal Pr ocessing (ICASSP) . IEEE, 2018. [11] J. W ang, J. Chen, D. Su, L. Chen, M. Y u, Y . Qian, and D. Y u, “Deep extractor network for target speaker recovery from sin- gle channel speech mixtures, ” in Pr oc. Interspeech , 2018. [12] L. Drude, T . von Neumann, and R. Haeb-Umbach, “Deep at- tractor networks for speaker re-identification and blind source separation, ” in 2018 IEEE International Conference on Acous- tics, Speech and Signal Processing (ICASSP) . IEEE, 2018. [13] P .-S. Huang, M. Kim, M. Hasegaw a-Johnson, and P . Smaragdis, “Singing-v oice separation from monaural recordings using deep recurrent neural networks. ” in ISMIR , 2014. [14] A. A. Nugraha, A. Liutkus, and E. V incent, “Multichannel mu- sic separation with deep neural networks, ” in Signal Processing Confer ence (EUSIPCO), 2016 24th Eur opean . IEEE, 2016. [15] S. Uhlich, M. Porcu, F . Giron, M. Enenkl, T . Kemp, N. T aka- hashi, and Y . Mitsufuji, “Improving music source separation based on deep neural networks through data augmentation and network blending, ” in Acoustics, Speech and Signal Pr ocessing (ICASSP), 2017 IEEE International Confer ence on . IEEE, 2017. [16] F .-R. St ¨ oter , A. Liutkus, and N. Ito, “The 2018 signal sepa- ration e valuation campaign, ” in International Confer ence on Latent V ariable Analysis and Signal Separation . Springer , 2018. [17] R. Kumar , Y . Luo, and N. Mesgarani, “Music source activity detection and separation using deep attractor network, ” Pr oc. Interspeech 2018 , 2018. [18] B. Pardo, “Finding structure in audio for music information retriev al, ” IEEE Signal Processing Ma gazine , vol. 23, no. 3, 2006. [19] D. El Badawy , N. Q. Duong, and A. Ozero v , “On-the-fly audio source separationa novel user-friendly framew ork, ” IEEE/ACM T ransactions on Audio, Speech, and Language Processing , vol. 25, no. 2, 2017. [20] A. Ozerov , S. Kiti ´ c, and P . P ´ erez, “ A comparative study of example-guided audio source separation approaches based on nonnegati ve matrix factorization, ” in Machine Learning for Signal Pr ocessing (MLSP), 2017 IEEE 27th International W orkshop on . IEEE, 2017. [21] B. McFee, J. Salamon, and J. P . Bello, “ Adapti ve pooling op- erators for weakly labeled sound event detection, ” IEEE/ACM T rans. on Audio, Speech, and Language Processing , vol. 26, no. 11, Nov . 2018. [22] Z. Rafii, A. Liutkus, F .-R. St ¨ oter , S. I. Mimilakis, and R. Bittner , “The MUSDB18 corpus for music separation, ” Dec. 2017. [23] J. Salamon, D. MacConnell, M. Cartwright, P . Li, and J. P . Bello, “Scaper: A library for soundscape synthesis and aug- mentation, ” in Applications of Signal Pr ocessing to Audio and Acoustics (W ASP AA), 2017 IEEE W orkshop on . IEEE, 2017.
Original Paper
Loading high-quality paper...
Comments & Academic Discussion
Loading comments...
Leave a Comment