Access the full text.
Sign up today, get DeepDyve free for 14 days.
(2003)
Significance of source features for speaker recognition, Master’s Thesis, MS Thesis, Department of Computer Science and Engineering, Indian Institute of Technology
Krothapalli Rao, B. Yegnanarayana (2009)
Intonation modeling for Indian languages
Srinivas Desai, A. Black, B. Yegnanarayana, K. Prahallad (2010)
Spectral Mapping Using Artificial Neural Networks for Voice ConversionIEEE Transactions on Audio, Speech, and Language Processing, 18
S. Gangashetty, C. Sekhar, B. Yegnanarayana (2004)
Detection of vowel on set points in continuous speech using autoassociative neural network models
(2000)
Autoassociate neural network models for processing degraded speech
S. Prasanna, B. Reddy, P. Krishnamoorthy (2009)
Vowel Onset Point Detection Using Source, Spectral Peaks, and Modulation Spectrum EnergiesIEEE Transactions on Audio, Speech, and Language Processing, 17
Yumin Zeng, Huayu Wu, Rongchun Gao (2007)
Pitch Synchronous Analysis Method and Fisher Criterion Based Speaker IdentificationThird International Conference on Natural Computation (ICNC 2007), 2
S. Koolagudi, Sreenivasa Krothapalli (2012)
Emotion recognition from speech using sub-syllabic and pitch synchronous spectral featuresInternational Journal of Speech Technology, 15
K. Rao, B. Yegnanarayana (2009)
Duration modification using glottal closure instants and vowel onset pointsSpeech Commun., 51
(2004)
Pattern Classification, 2nd edn. (A Wiley-interscience Publications, Singapore
S. Ikbal, Hemant Misra, B. Yegnanarayana (1999)
Analysis of autoassociative mapping neural networksIJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339), 5
S. Haykin (1998)
Neural Networks: A Comprehensive Foundation
S. Kishore, B. Yegnanarayana, S. Gangashetty (2001)
Online text-independent speaker verification system using autoassociative neural network modelsIJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222), 2
Hiroshi Muta, Thomas Baer, Kikuju Wagatsuma, T. Muraoka, H. Fukuda (1988)
A pitch-synchronous analysis of hoarseness in running speech.The Journal of the Acoustical Society of America, 84 4
(2004)
Source and system features for speaker recognition, Master’s Thesis, MS Thesis, Department of Computer Science and Engineering, Indian Institute of Technology
L. Mary, B. Yegnanarayana (2004)
Autoassociative neural network models for language identificationInternational Conference on Intelligent Sensing and Information Processing, 2004. Proceedings of
K. Murty, B. Yegnanarayana (2008)
Epoch Extraction From Speech SignalsIEEE Transactions on Audio, Speech, and Language Processing, 16
O. Mubarak, E. Ambikairajah, J. Epps (2005)
Analysis of an MFCC-based audio indexing system for efficient coding of multimedia sourcesProceedings of the Eighth International Symposium on Signal Processing and Its Applications, 2005., 2
(2003)
Significance of source features for speaker recognition
C. Mohan, B. Yegnanarayana, Krishna Mohan, B Yegnanarayana (2010)
Classification of sport videos using edge-based features and autoassociative neural network modelsSignal, Image and Video Processing, 4
B. Yegnanarayana, K. Reddy, K. Prahallad (2001)
Source and system features for speaker recognition using AANN models2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221), 1
F. Burkhardt, A. Paeschke, M. Rolfes, W. Sendlmeier, Benjamin Weiss (2005)
A database of German emotional speech
(2006)
Resources, features, and methods
L. Rabiner, B. Juang (1993)
Fundamentals of speech recognition
D. Bitouk, R. Verma, A. Nenkova (2010)
Class-level spectral features for emotion recognitionSpeech communication, 52 7-8
(2004)
Detection of vowel on set points in continuous speech using auto-associative neural network models, in INTERSPEECH
(2000)
Autoassociate neural network models for processing degraded speech, Master’s Thesis, MS Thesis, Department of Computer Science and Engineering, Indian Institute of Technology
(2001)
Kishore, Source and system features for speaker recognition using aann models, in IEEE International Conference on Acoustics, Speech, and Signal Processing, (Salt Lake City, UT)
(2004)
Pattern Classification, 2nd edn
John Sinclair, Herbert Wiegand, Richard Allsopp, James Arthurs, Arthur Bronstein, Louise Dagenais, Robert London, Christian Kay, Johan Leiden, Franciscus Junius, B. Kipfer, Alan Kirkness, A. Cowie, Ton Broeders, Phil Hyams, Thomas Creamer, Takoma Park, N. Kharma, Hans-Peder Kromann, Theis Riiber, Poul Rosbach, E. Lovatt, J. Mdee, H. Niedzielski, Viggo Pedersen, Sandra Thompson (1992)
TABLE OF CONTENTS Preface
S. Koolagudi, Sudhamay Maity, A. Vuppala, S. Chakrabarti, K. Rao (2009)
IITKGP-SESC: Speech Database for Emotion Analysis
D. Ververidis, Constantine Kotropoulos (2006)
Emotional speech recognition: Resources, features, and methodsSpeech Commun., 48
B. Yegnanarayana, K. Prahallad (2002)
AANN: an alternative to GMM for pattern recognitionNeural networks : the official journal of the International Neural Network Society, 15 3
C. Gupta, S. Prasanna, B. Yegnanarayana (2002)
Autoassociative neural network models for online speaker verification using source features from vowelsProceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290), 2
Jingdong Chen, Yiteng Huang, Qi Li, K. Paliwal (2004)
Recognition of noisy speech using dynamic spectral subband centroidsIEEE Signal Processing Letters, 11
(1993)
Fundamentals of Speech Recognition (Prentice-Hall, Englewood Cliffs, New Jersy
D. Neiberg, K. Elenius, K. Laskowski (2006)
Emotion recognition in spontaneous speech using GMMs
(1999)
Neural Networks: A Comprehensive Foundation (Pearson Education Aisa, Inc
[This chapter discusses the use of vocal tract information for recognizing the emotions. Linear prediction cepstral coefficients (LPCC) and mel frequency cepstral coefficients (MFCC) are used as the correlates of vocal tract information. In addition to LPCCs and MFCCs, formant related features are also explored in this work for recognizing emotions from speech. Extraction of the above mentioned spectral features is discussed in brief. Further extraction of these features from sub-syllabic regions such as consonants, vowels and consonant-vowel transition regions is discussed. Extraction of spectral features from pitch synchronous analysis is also discussed. Basic philosophy and use of Gaussian mixture models is discussed in this chapter for classifying the emotions. The emotion recognition performance obtained from different vocal tract features is compared. Proposed spectral features are evaluated on Indian and Berlin emotion databases. Performance of Gaussian mixture models in classifying the emotional utterances using vocal tract features is compared with neural network models.]
Published: Jan 13, 2013
Keywords: Speech Signal; Emotion Recognition; Vocal Tract; Speech Emotion Recognition; Emotion Recognition System
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.