Papers
Topics
Authors
Recent
Search
2000 character limit reached

Instabilities in Convnets for Raw Audio

Published 11 Sep 2023 in cs.LG, cs.SD, and eess.AS | (2309.05855v4)

Abstract: What makes waveform-based deep learning so hard? Despite numerous attempts at training convolutional neural networks (convnets) for filterbank design, they often fail to outperform hand-crafted baselines. These baselines are linear time-invariant systems: as such, they can be approximated by convnets with wide receptive fields. Yet, in practice, gradient-based optimization leads to suboptimal approximations. In our article, we approach this phenomenon from the perspective of initialization. We present a theory of large deviations for the energy response of FIR filterbanks with random Gaussian weights. We find that deviations worsen for large filters and locally periodic input signals, which are both typical for audio signal processing applications. Numerical simulations align with our theory and suggest that the condition number of a convolutional layer follows a logarithmic scaling law between the number and length of the filters, which is reminiscent of discrete wavelet bases.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. T. Necciari, N. Holighaus, P. Balazs, Z. Průša, P. Majdak, and O. Derrien, “Audlet filter banks: A versatile analysis/synthesis framework using auditory frequency scales,” Applied Sciences, vol. 8, no. 1, 2018.
  2. M.-A. Meier, T. Heaton, and J. Clinton, “The Gutenberg algorithm: Evolutionary Bayesian magnitude estimates for earthquake early warning with a filter bank,” Bulletin of the Seismological Society of America, vol. 105, no. 5, pp. 2774–2786, 2015.
  3. E. Chassande-Mottin, “Learning approach to the detection of gravitational wave transients,” Physical Review D, vol. 67, no. 10, 2003.
  4. K. K. Ang, Z. Y. Chin, H. Zhang, and C. Guan, “Filter bank common spatial pattern (FBCSP) in brain-computer interface,” in Proc. IEEE IJCNN, 2008, pp. 2390–2397.
  5. M. Dörfler, T. Grill, R. Bammer, and A. Flexer, “Basic filters for convolutional neural networks applied to music: Training or design?” Neural Computing and Applications, vol. 32, pp. 941–954, 2020.
  6. M. Ravanelli and Y. Bengio, “Speaker recognition from raw waveform with SincNet,” in Proc. IEEE SLT, 2018, pp. 1021–1028.
  7. N. Zeghidour, O. Teboul, F. de Chaumont Quitry, and M. Tagliasacchi, “LEAF: A learnable frontend for audio classification,” in Proc. ICLR, 2021.
  8. N. Zeghidour, N. Usunier, I. Kokkinos, T. Schaiz, G. Synnaeve, and E. Dupoux, “Learning filterbanks from raw speech for phone recognition,” in Proc. IEEE ICASSP, 2018, pp. 5509–5513.
  9. T. Sainath, R. J. Weiss, K. Wilson, A. W. Senior, and O. Vinyals, “Learning the speech front-end with raw waveform CLDNNs,” in Proc. INTERSPEECH, 2015.
  10. D. Stowell, “Computational bioacoustics with deep learning: A review and roadmap,” PeerJ, vol. 10, 2022.
  11. I. López-Espejo, Z.-H. Tan, and J. Jensen, “Exploring filterbank learning for keyword spotting,” in Proc. EUSIPCO, 2021, pp. 331–335.
  12. J. Schlüter and G. Gutenbrunner, “EfficientLEAF: A faster learnable audio frontend of questionable use,” in Proc. EUSIPCO.   IEEE, 2022, pp. 205–208.
  13. F. J. Bravo Sanchez, M. R. Hossain, N. B. English, and S. T. Moore, “Bioacoustic classification of avian calls from raw sound waveforms with an open-source deep learning architecture,” Scientific Reports, vol. 11, no. 1, pp. 1–12, 2021.
  14. V. Lostanlen, D. Haider, H. Han, M. Lagrange, P. Balazs, and M. Ehler, “Fitting auditory filterbanks with multiresolution neural networks,” in Proc. IEEE WASPAA, 2023, pp. 1–5.
  15. M. Cisse, P. Bojanowski, E. Grave, Y. Dauphin, and N. Usunier, “Parseval networks: Improving robustness to adversarial examples,” Proc. ICML, 2017.
  16. K. Gupta, F. Kaakai, B. Pesquet-Popescu, J.-C. Pesquet, and F. D. Malliaros, “Multivariate Lipschitz analysis of the stability of neural networks,” Frontiers in Signal Processing, 2022.
  17. A. Virmaux and K. Scaman, “Lipschitz regularity of deep neural networks: Analysis and efficient estimation,” Advances in Neural Information Processing Systems, vol. 31, 2018.
  18. M. Ehler, “Preconditioning filter bank decomposition using structured normalized tight frames,” Journal of Applied Mathematics, vol. 2015, pp. 1 – 12, 2015.
  19. H. Chernoff, “A measure of asymptotic efficiency for tests of a hypothesis based on the sum of observations,” The Annals of Mathematical Statistics, vol. 23, no. 4, pp. 493 – 507, 1952.
  20. L. Birgé and P. Massart, “Minimum contrast estimators on sieves: exponential bounds and rates of convergence,” Bernoulli, vol. 4, no. 3, pp. 329–375, 1998.
  21. B. Laurent and P. Massart, “Adaptive estimation of a quadratic functional by model selection,” Annals of Statistics, vol. 28, 2000.
  22. G. Barrera Vargas and P. Manrique-Mirón, “The asymptotic distribution of the condition number for random circulant matrices,” Extremes, vol. 25, 2022.
Citations (1)

Summary

  • The paper reveals that initialization issues cause large deviations in energy response of CNN filters, undermining raw audio processing stability.
  • It demonstrates that the convolutional layer’s condition number scales logarithmically with filter count and length, affecting network performance.
  • The study underscores the need for robust initialization strategies to close the performance gap between CNN approaches and hand-crafted filterbanks.

The paper "Instabilities in Convnets for Raw Audio" (2309.05855) explores the complexities of training convolutional neural networks (CNNs) specifically for raw audio applications. The core issue addressed is why CNNs often struggle to outperform hand-crafted baselines in tasks such as filterbank design, even though these baselines are linear time-invariant systems that could theoretically be emulated by CNNs with wide receptive fields.

The authors approach this problem from the perspective of initialization and present a theoretical framework for understanding large deviations in the energy response of finite impulse response (FIR) filterbanks with random Gaussian weights. They identify that these deviations become more problematic for larger filters and for input signals that are locally periodic—both common characteristics in audio signal processing. Their findings suggest that the condition number of a convolutional layer exhibits a logarithmic scaling relationship with both the number and length of filters, echoing properties seen in discrete wavelet bases.

For context, modeling raw audio waveforms has been an active research area, with various approaches explored in the literature:

  1. WaveNet: A probabilistic, autoregressive model for raw audio that has shown remarkable performance in text-to-speech applications (Oord et al., 2016).
  2. SaShiMi: A multi-scale architecture leveraging state-space models, which addresses instabilities in autoregressive generation for raw audio (Goel et al., 2022).
  3. Very Deep CNNs: These aim to use extensive convolutional layers to capture higher-level features directly from time-domain waveforms (Dai et al., 2016).

In addition to these, techniques like FloWaveNet have been proposed to mitigate the inefficiencies of ancestral sampling in models like WaveNet, providing more efficient real-time audio synthesis by leveraging flow-based generative models (Kim et al., 2018).

In summary, while there are ongoing advancements in modeling raw audio with neural networks, the paper highlights inherent initialization-related instabilities as a critical challenge for CNN-based approaches in this domain. This insight helps frame future research directions aimed at creating more robust and effective models for raw audio processing.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 4 tweets with 33 likes about this paper.