Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/16474
Title: A maximum log-likelihood approach to voice activity detection
Authors: Gauci, Oliver
Debono, Carl James
Micallef, Paul
Keywords: Speech processing systems
Random noise theory
Ambient sounds
Issue Date: 2008
Publisher: Institute of Electrical and Electronics Engineers Inc.
Citation: Gauci, O., Debono, C. J., & Micallef, P. (2008). A maximum log-likelihood approach to voice activity detection. 3rd International Symposium on Communications, Control and Signal Processing, St. Julians. 383-387.
Abstract: Modern voice activity detection (VAD) algorithms must achieve reliable operation at low signal-to-noise ratios (SNR). Although a lot of research has been performed to solve this issue, the operation of existing VAD algorithms is still far away from ideal. In this paper, we present a novel VAD algorithm, in which we apply the Teager energy cepstral coefficients, to obtain a noise robust feature extraction method, together with Gaussian mixture models that serve for the classification of speech and silence periods. In the suggested solution, the threshold method used in many noise robust VAD algorithms is eliminated, thus favoring its use in real applications. The performance of this novel algorithm was tested under known and unknown noise statistics, and compared to a statistical model-based approach found in literature. The results obtained show that the proposed solution achieves better accuracy and significantly reduces clipping of speech periods; thus achieving superior signal quality.
URI: https://www.um.edu.mt/library/oar//handle/123456789/16474
Appears in Collections:Scholarly Works - FacICTCCE

Files in This Item:
File Description SizeFormat 
Conference Paper - A maximum log-likelihood approach to voice activity detection.pdf
  Restricted Access
A maximum log-likelihood approach to voice activity detection309.87 kBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.