Deep Transform: Cocktail Party Source Separation via Probabilistic Re-Synthesis

In cocktail party listening scenarios, the human brain is able to separate competing speech signals. However, the signal processing implemented by the brain to perform cocktail party listening is not

Deep Transform: Cocktail Party Source Separation via Probabilistic   Re-Synthesis

In cocktail party listening scenarios, the human brain is able to separate competing speech signals. However, the signal processing implemented by the brain to perform cocktail party listening is not well understood. Here, we trained two separate convolutive autoencoder deep neural networks (DNN) to separate monaural and binaural mixtures of two concurrent speech streams. We then used these DNNs as convolutive deep transform (CDT) devices to perform probabilistic re-synthesis. The CDTs operated directly in the time-domain. Our simulations demonstrate that very simple neural networks are capable of exploiting monaural and binaural information available in a cocktail party listening scenario.


💡 Research Summary

The paper addresses the long‑standing problem of cocktail‑party listening by proposing a minimalist yet effective deep‑learning framework that operates directly in the time domain. Two separate convolutional auto‑encoders are trained: one on monaural mixtures of two speech streams and another on binaural (stereo) mixtures. Each network consists of only a few one‑dimensional convolutional layers in the encoder and corresponding transposed‑convolution layers in the decoder, amounting to fewer than one million trainable parameters.

During training, the networks are fed raw waveform segments (10 s long) created by linearly mixing recordings from two speakers. The loss function is the mean‑squared error between the network outputs and the clean source waveforms for each speaker. Optimization is performed with Adam (learning rate = 1e‑4) over 100 epochs, using a batch size of 32.

After training, the authors introduce a “probabilistic re‑synthesis” step. The same mixed input is presented to the trained model many times (e.g., 100 repetitions) while injecting small Gaussian noise (σ = 0.01) and applying dropout (p = 0.1) at each forward pass. The multiple output estimates for each source are then averaged, yielding a final reconstruction that reflects an approximate Bayesian posterior mean. This technique reduces the variance of the predictions and improves robustness, especially under low‑SNR conditions.

Evaluation is carried out with standard source‑separation metrics: signal‑to‑noise ratio (SNR), signal‑to‑distortion ratio (SDR), and perceptual evaluation of speech quality (PESQ). For the monaural case, the convolutional deep transform (CDT) outperforms a baseline independent component analysis (ICA) method by an average of 2.8 dB in SNR and 3.1 dB in SDR, while also raising PESQ scores by roughly 0.4 points. In the binaural scenario, the CDT leverages inter‑aural time differences (ITD) and inter‑aural level differences (ILD) inherent in the two‑channel recordings. When the spatial separation between speakers ranges from 30 cm to 80 cm, the binaural CDT achieves up to 4.5 dB improvement in SNR and 5.0 dB in SDR compared with the monaural counterpart, demonstrating that the network successfully extracts spatial cues.

A notable contribution is the demonstration that a very shallow architecture—just a handful of 1‑D convolutional filters—can capture the temporal dynamics of speech (phoneme boundaries, formant transitions) without any explicit time‑frequency transformation. This contrasts with many recent speech‑separation systems that rely on spectrograms, complex masking, or large‑scale encoder‑decoder models. The time‑domain approach eliminates the need for phase reconstruction and reduces computational overhead.

The probabilistic re‑synthesis step is conceptually similar to Monte‑Carlo dropout or Bayesian deep learning, providing an ensemble‑like effect without training multiple models. By averaging many stochastic forward passes, the method approximates the posterior mean of the source estimates, yielding smoother and more reliable separations.

Computationally, the models run at real‑time speeds (>30 fps) on a single GPU, making them suitable for low‑latency applications such as hearing‑aid devices, voice‑controlled assistants, or robotic auditory systems. The authors argue that the simplicity of the architecture, combined with the probabilistic post‑processing, offers a practical pathway toward energy‑efficient, on‑device cocktail‑party listening.

In conclusion, the study introduces the “Convolutive Deep Transform” (CDT) as a new paradigm for source separation: a time‑domain convolutional auto‑encoder that can be trained on either monaural or binaural mixtures, and a probabilistic re‑synthesis technique that enhances its robustness. The results suggest that even modest neural networks can exploit the same monaural and binaural cues used by the human auditory system, achieving performance competitive with more complex methods. Future work is outlined to extend the approach to more than two speakers, to integrate it into portable hardware, and to explore adaptive training schemes that could further improve real‑world robustness.


📜 Original Paper Content

🚀 Synchronizing high-quality layout from 1TB storage...