Skip to main content

An FFT-Based Companding Front End for Noise-Robust Automatic Speech Recognition

Abstract

We describe an FFT-based companding algorithm for preprocessing speech before recognition. The algorithm mimics tone-to-tone suppression and masking in the auditory system to improve automatic speech recognition performance in noise. Moreover, it is also very computationally efficient and suited to digital implementations due to its use of the FFT. In an automotive digits recognition task with the CU-Move database recorded in real environmental noise, the algorithm improves the relative word error by 12.5% at -5 dB signal-to-noise ratio (SNR) and by 6.2% across all SNRs (-5 dB SNR to +5 dB SNR). In the Aurora-2 database recorded with artificially added noise in several environments, the algorithm improves the relative word error rate in almost all situations.

[1234567891011121314151617181920212223242526272829303132]

References

  1. 1.

    Lippmann RP: Speech recognition by machines and humans. Speech Communication 1997,22(1):1-15. 10.1016/S0167-6393(97)00021-6

    Article  Google Scholar 

  2. 2.

    Pickles JO: An Introduction to the Physiology of Hearing. Academic Press, London, UK; 1988.

    Google Scholar 

  3. 3.

    Seneff S: A joint synchrony/mean-rate model of auditory speech processing. Journal of Phonetics 1988,16(1):55-76.

    Google Scholar 

  4. 4.

    Ghitza O: Auditory models and human performance in tasks related to speech coding and speech recognition. IEEE Transactions on Speech and Audio Processing 1994,2(1, part 2):115-132. 10.1109/89.260357

    Article  Google Scholar 

  5. 5.

    Van Schaik A, Meddis R: Analog very large-scale integrated (VLSI) implementation of a model of amplitude-modulation sensitivity in the auditory brainstem. Journal of the Acoustical Society of America 1999,105(2):811-821. 10.1121/1.426270

    Article  Google Scholar 

  6. 6.

    Goldstein JL: Modeling rapid waveform compression on the basilar membrane as multiple-bandpass-nonlinearity filtering. Hearing Research 1990,49(1–3):39-60.

    Article  Google Scholar 

  7. 7.

    Meddis R, O'Mard LP, Lopez-Poveda EA: A computational algorithm for computing nonlinear auditory frequency selectivity. Journal of the Acoustical Society of America 2001,109(6):2852-2861. 10.1121/1.1370357

    Article  Google Scholar 

  8. 8.

    Jankowski CR Jr., Vo H-DH, Lippmann RP: A comparison of signal processing front ends for automatic word recognition. IEEE Transactions on Speech and Audio Processing 1995,3(4):286-293. 10.1109/89.397093

    Article  Google Scholar 

  9. 9.

    Davis SB, Mermelstein P: Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Transactions on Acoustics, Speech, and Signal Processing 1980,28(4):357-366. 10.1109/TASSP.1980.1163420

    Article  Google Scholar 

  10. 10.

    Hermansky H: Perceptual linear predictive (PLP) analysis of speech. Journal of the Acoustical Society of America 1990,87(4):1738-1752. 10.1121/1.399423

    Article  Google Scholar 

  11. 11.

    Strope B, Alwan A: A model of dynamic auditory perception and its application to robust word recognition. IEEE Transactions on Speech and Audio Processing 1997,5(5):451-464. 10.1109/89.622569

    Article  Google Scholar 

  12. 12.

    Holmberg M, Gelbart D, Hemmert W: Automatic speech recognition with an adaptation model motivated by auditory processing. IEEE Transactions on Audio, Speech, and Language Processing 2006,14(1):43-49.

    Article  Google Scholar 

  13. 13.

    Tchorz J, Kollmeier B: A model of auditory perception as front end for automatic speech recognition. Journal of the Acoustical Society of America 1999,106(4):2040-2050. 10.1121/1.427950

    Article  Google Scholar 

  14. 14.

    Hermansky H, Morgan N: RASTA processing of speech. IEEE Transactions on Speech and Audio Processing 1994,2(4):578-589. 10.1109/89.326616

    Article  Google Scholar 

  15. 15.

    Moore BCJ: An Introduction to the Psychology of Hearing. 4th edition. Academic Press, New York, NY, USA; 1997.

    Google Scholar 

  16. 16.

    Turicchia L, Sarpeshkar R: The silicon cochlea: from biology to bionics. In Biophysics of the Cochlea: From Molecules to Models. Edited by: Gummer AW. World Scientific, Singapore; 2003:417-423.

    Google Scholar 

  17. 17.

    Turicchia L, Sarpeshkar R: A bio-inspired companding strategy for spectral enhancement. IEEE Transactions on Speech and Audio Processing 2005,13(2):243-253.

    Article  Google Scholar 

  18. 18.

    Oxenham AJ, Simonson AM, Turicchia L, Sarpeshkar R: Evaluation of companding-based spectral enhancement using simulated cochlear-implant processing. Journal of the Acoustical Society of America 2007,121(3):1709-1716. 10.1121/1.2434757

    Article  Google Scholar 

  19. 19.

    Bhattacharya A, Zeng F-G: Companding to improve cochlear implants' speech processing in noise. Proceedings of Conference on Implantable Auditory Prostheses, July-August 2005, Pacific Grove, Calif, USA

    Google Scholar 

  20. 20.

    Lee YW, Kwon SY, Ji YS, et al.: Speech enhancement in noise environment using companding strategy. Proceedings of the 5th Asia Pacific Symposium on Cochlear Implant and Related Sciences (APSCI '05), November 2005, Hong Kong

    Google Scholar 

  21. 21.

    Loizou PC, Kasturi K, Turicchia L, Sarpeshkar R, Dorman M, Spahr T: Evaluation of the companding and other strategies for noise reduction in cochlear implants. Proceedings of Conference on Implantable Auditory Prostheses, July-August 2005, Pacific Grove, Calif, USA

    Google Scholar 

  22. 22.

    Turicchia L, Kasturi K, Loizou PC, Sarpeshkar R: Evaluation of the companding algorithm for noise reduction in cochlear implants. submitted for publication

  23. 23.

    Guinness J, Raj B, Schmidt-Nielsen B, Turicchia L, Sarpeshkar R: A companding front end for noise-robust automatic speech recognition. Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '05), March 2005, Philadelphia, Pa, USA 1: 249-252.

    Google Scholar 

  24. 24.

    Stone MA, Moore BCJ: Spectral feature enhancement for people with sensorineural hearing impairment: effects on speech intelligibility and quality. Journal of Rehabilitation Research and Development 1992,29(2):39-56. 10.1682/JRRD.1992.04.0039

    Article  Google Scholar 

  25. 25.

    Baer T, Moore BCJ, Gatehouse S: Spectral contrast enhancement of speech in noise for listeners with sensorineural hearing impairment: effects on intelligibility, quality, and response times. Journal of Rehabilitation Research and Development 1993,30(1):49-72.

    Google Scholar 

  26. 26.

    Juang B-H, Rabiner LR, Wilpon JG: On the use of bandpass liftering in speech recognition. IEEE Transactions on Acoustics, Speech, and Signal Processing 1987,35(7):947-954. 10.1109/TASSP.1987.1165237

    Article  Google Scholar 

  27. 27.

    Hunt MJ: Some experience in in-car speech recognition. Proceedings of the Workshop on Robust Methods for Speech Recognition in Adverse Conditions, May 1999, Tampere, Finland 25-31.

    Google Scholar 

  28. 28.

    University Technology Corporation : CSLR Speech Corpora. http://cslr.colorado.edu/beginweb/speechcorpora/corpus.html

  29. 29.

    Hirsch H-G, Pearce D: The AURORA experimental framework for the performance evaluation of speech recognition systems under noisy conditions. Proceedings of Automatic Speech Recognition: Challenges for the New Millenium (ISCA ITRW ASR '00), September 2000, Paris, France 181-188.

    Google Scholar 

  30. 30.

    Kandel ER, Schwarz JH, Jessell TM: Principles of Neural Science. McGraw Hill, New York, NY, USA; 2000.

    Google Scholar 

  31. 31.

    Singh R, Seltzer ML, Raj B, Stern RM: Speech in noisy environments: robust automatic segmentation, feature extraction, and hypothesis combination. Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '01), May 2001, Salt Lake, Utah, USA 1: 273-276.

    Google Scholar 

  32. 32.

    The Hidden Markov Model Toolkit (HTK) University of Cambridge, http://htk.eng.cam.ac.uk/

Download references

Author information

Affiliations

Authors

Corresponding author

Correspondence to Bhiksha Raj.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Raj, B., Turicchia, L., Schmidt-Nielsen, B. et al. An FFT-Based Companding Front End for Noise-Robust Automatic Speech Recognition. J AUDIO SPEECH MUSIC PROC. 2007, 065420 (2007). https://doi.org/10.1155/2007/65420

Download citation

Keywords

  • Error Rate
  • Acoustics
  • Recognition Task
  • Recognition Performance
  • Auditory System