Robust Classification of Stop Consonants Using Auditory-Based Speech Processing

Loading...
Thumbnail Image
Penn collection
Departmental Papers (ESE)
Degree type
Discipline
Subject
features
stop consonants
speaker independent speech
speech recognition
auditory-based speech processing
acoutic-phonetic feature
average localized synchrony
ALSD
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Abdelatty Ali, Ahmed M
Mueller, Paul
Contributor
Abstract

In this work, a feature-based system for the automatic classification of stop consonants, in speaker independent continuous speech, is reported. The system uses a new auditory-based speech processing front-end that is based on the biologically rooted property of average localized synchrony detection (ALSD). It incorporates new algorithms for the extraction and manipulation of the acoustic-phonetic features that proved, statistically, to be rich in their information content. The experiments are performed on stop consonants extracted from the TIMIT database with additive white Gaussian noise at various signal-to-noise ratios. The obtained classification accuracy compares favorably with previous work. The results also showed a consistent improvement of 3% in the place detection over the Generalized Synchrony Detector (GSD) system under identical circumstances on clean and noisy speech. This illustrates the superior ability of the ALSD to suppress the spurious peaks and produce a consistent and robust formant (peak) representation.

Advisor
Date of presentation
2001-05-07
Conference name
Departmental Papers (ESE)
Conference dates
2023-05-16T21:43:53.000
Conference location
Date Range for Data Collection (Start Date)
Date Range for Data Collection (End Date)
Digital Object Identifier
Series name and number
Volume number
Issue number
Publisher
Publisher DOI
Journal Issue
Comments
Copyright 2001 IEEE. Reprinted from Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing 2001 (ICASSP 2001) Volume 1, pages 81-84. This material is posted here with permission of the IEEE. Such permission of the IEEE does not in any way imply IEEE endorsement of any of the University of Pennsylvania's products or services. Internal or personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution must be obtained from the IEEE by writing to pubs-permissions@ieee.org. By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
Copyright 2001 IEEE. Reprinted from Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing 2001 (ICASSP 2001) Volume 1, pages 81-84. Publisher URL: http://ieeexplore.ieee.org/xpl/tocresult.jsp?isNumber=20365&page=1 This material is posted here with permission of the IEEE. Such permission of the IEEE does not in any way imply IEEE endorsement of any of the University of Pennsylvania's products or services. Internal or personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution must be obtained from the IEEE by writing to pubs-permissions@ieee.org. By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
Recommended citation
Collection