메뉴 건너뛰기




Volumn 15, Issue 2, 2012, Pages 265-289

Emotion recognition from speech using source, system, and prosodic features

Author keywords

Emo DB; Emotion recognition; Excitation source features; Global prosodic features; Glottal closure instants; IITKGP SESC; Local prosodic features; Pitch synchronous analysis; System features; Zero frequency filter

Indexed keywords

EMO-DB; EMOTION RECOGNITION; EXCITATION SOURCES; GLOTTAL CLOSURE INSTANTS; IITKGP-SESC; PITCH SYNCHRONOUS ANALYSIS; PROSODIC FEATURES; SYSTEM FEATURES; ZERO FREQUENCY;

EID: 84864708818     PISSN: 13812416     EISSN: 15728110     Source Type: Journal    
DOI: 10.1007/s10772-012-9139-3     Document Type: Article
Times cited : (77)

References (74)
  • 1
    • 84856275800 scopus 로고    scopus 로고
    • Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India
    • Anjani, A. V. N. S. (2000). Autoassociate neural network models for processing degraded speech.Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India.
    • (2000) Autoassociate neural network models for processing degraded speech
    • Anjani, A.V.N.S.1
  • 2
    • 0015476226 scopus 로고
    • Automatic speaker recognition based on pitch contours
    • Atal, B. S. (1972). Automatic speaker recognition based on pitch contours. The Journal of the Acoustical Society of America, 52(6), 1687-1697.
    • (1972) The Journal of the Acoustical Society of America , vol.52 , Issue.6 , pp. 1687-1697
    • Atal, B.S.1
  • 4
    • 21844456055 scopus 로고    scopus 로고
    • The role of intonation in emotional expressions
    • DOI 10.1016/j.specom.2005.02.016, PII S0167639305000890, Quantitative Prosody Modelling for Natural Speech Description and Generation
    • Banziger, T., & Scherer, K. R. (2005). The role of intonation in emotional expressions. Speech Communication, 46, 252-267. (Pubitemid 40952515)
    • (2005) Speech Communication , vol.46 , Issue.3-4 , pp. 252-267
    • Banziger, T.1    Scherer, K.R.2
  • 6
    • 77956401353 scopus 로고    scopus 로고
    • Class-level spectral features for emotion recognition
    • Bitouk, D., Verma, R., & Nenkova, A. (2010). Class-level spectral features for emotion recognition. Speech Communication, 52(7-8), 613-625.
    • (2010) Speech Communication , vol.52 , Issue.7-8 , pp. 613-625
    • Bitouk, D.1    Verma, R.2    Nenkova, A.3
  • 7
    • 0002689942 scopus 로고    scopus 로고
    • Verification of acoustical correlates of emotional speech using formant synthesis
    • Newcastle, Northern Ireland, UK, Sept. 2000
    • Burkhardt, F., & Sendlmeier, W. F. (2000). Verification of acoustical correlates of emotional speech using formant synthesis. In ITRW on speech and emotion, Newcastle, Northern Ireland, UK, Sept. 2000 (pp. 151-156).
    • (2000) ITRW on speech and emotion , pp. 151-156
    • Burkhardt, F.1    Sendlmeier, W.F.2
  • 9
    • 0002515370 scopus 로고
    • The generation of affect in synthesized speech
    • Jul. 1990
    • Cahn, J. E. (1990). The generation of affect in synthesized speech. In JAVIOS, Jul. 1990 (pp. 1-19).
    • (1990) JAVIOS , pp. 1-19
    • Cahn, J.E.1
  • 10
    • 0037382510 scopus 로고    scopus 로고
    • Describing the emotional states that are expressed in speech
    • Cowie, R., & Cornelius, R. R. (2003). Describing the emotional states that are expressed in speech. Speech Communication, 40, 5-32.
    • (2003) Speech Communication , vol.40 , pp. 5-32
    • Cowie, R.1    Cornelius, R.R.2
  • 12
    • 77958460688 scopus 로고    scopus 로고
    • Recognising emotions in speech
    • 96, Oct. 1996.
    • Dellaert, F., Polzin, T., & Waibel, A. (1996). Recognising emotions in speech. In ICSLP 96, Oct. 1996.
    • (1996) ICSLP
    • Dellaert, F.1    Polzin, T.2    Waibel, A.3
  • 16
    • 84856240372 scopus 로고    scopus 로고
    • Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India.
    • Gupta, C. S. (2003). Significance of source features for speaker recognition. Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India.
    • (2003) Significance of Source Features for Speaker Recognition
    • Gupta, C.S.1
  • 17
    • 0036082789 scopus 로고    scopus 로고
    • Autoassociative neural network models for online speaker verification using source features from vowels
    • Honolulu, Hawaii, USA, May 2002.
    • Gupta, C. S., Prasanna, S. R. M., & Yegnanarayana, B. (2002). Autoassociative neural network models for online speaker verification using source features from vowels. In Int. joint conf. neural networks, Honolulu, Hawaii, USA, May 2002.
    • (2002) Int. joint conf. neural networks
    • Gupta, C.S.1    Prasanna, S.R.M.2    Yegnanarayana, B.3
  • 18
    • 44949264114 scopus 로고    scopus 로고
    • Feature analysis for emotion recognition from Mandarin speech considering the special characteristics of Chinese language
    • Pittsburgh, Pennsylvania, Sept. 2006
    • hao Kao, Y.,&shan Lee, L. (2006). Feature analysis for emotion recognition from Mandarin speech considering the special characteristics of Chinese language. In INTERSPEECH-ICSLP, Pittsburgh, Pennsylvania, Sept. 2006 (pp. 1814-1817).
    • (2006) INTERSPEECH-ICSLP , pp. 1814-1817
    • Hao Kao, Y.1    Shan Lee, L.2
  • 28
    • 70450182943 scopus 로고    scopus 로고
    • Analysis of laugh signals for detecting in continuous speech
    • Brighton, UK, September 6-10
    • Kumar, K. S., Reddy, M. S. H., Murty, K. S. R., & Yegnanarayana, B. (2009). Analysis of laugh signals for detecting in continuous speech. In INTERSPEECH-09, Brighton, UK, September 6-10 (pp. 1591-1594.
    • (2009) INTERSPEECH , pp. 1591-1594
    • Kumar, K.S.1    Reddy, M.S.H.2    Murty, K.S.R.3    Yegnanarayana, B.4
  • 29
    • 14644439843 scopus 로고    scopus 로고
    • Toward detecting emotions in spoken dialogs
    • DOI 10.1109/TSA.2004.838534
    • Lee, C. M., & Narayanan, S. S. (2005). Toward detecting emotions in spoken dialogs. IEEE Transactions on Speech and Audio Processing, 13, 293-303. (Pubitemid 40320247)
    • (2005) IEEE Transactions on Speech and Audio Processing , vol.13 , Issue.2 , pp. 293-303
    • Lee, C.M.1    Narayanan, S.S.2
  • 32
    • 34547496515 scopus 로고    scopus 로고
    • The relevance of voice quality features in speaker independent emotion recognition
    • DOI 10.1109/ICASSP.2007.367152, 4218026, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '07
    • Lugger, M., & Yang, B. (2007). The relevance of voice quality features in speaker independent emotion recognition. In ICASSP, Honolulu, Hawaii, USA, May 2007 (pp. IV17-IV20). New York: IEE (Pubitemid 47178301)
    • (2007) ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings , vol.4
    • Lugger, M.1    Yang, B.2
  • 35
    • 84884947988 scopus 로고    scopus 로고
    • Classification of sport videos using edge-based features and autoassociative neural network models
    • Mohan, C. K., & Yegnanarayana, B. (2008). Classification of sport videos using edge-based features and autoassociative neural network models. Signal, Image and Video Processing, 4, 61-73.
    • (2008) Signal, Image and Video Processing , vol.4 , pp. 61-73
    • Mohan, C.K.1    Yegnanarayana, B.2
  • 37
    • 0029325035 scopus 로고
    • Implementation and testing of a system for producing emotion by rule in synthetic speech
    • Murray, I. R., & Arnott, J. L. (1995). Implementation and testing of a system for producing emotion by rule in synthetic speech. Speech Communication, 16, 369-390.
    • (1995) Speech Communication , vol.16 , pp. 369-390
    • Murray, I.R.1    Arnott, J.L.2
  • 38
    • 0030291449 scopus 로고    scopus 로고
    • Emotional stress in synthetic speech: Progress and future directions
    • Murray, I. R., Arnott, J. L., & Rohwer, E. A. (1996). Emotional stress in synthetic speech: Progress and future directions. Speech Communication, 20, 85-91.
    • Speech Communication , vol.20 , pp. 85-91
    • Murray, I.R.1    Arnott, J.L.2    Rohwer, E.A.3
  • 41
    • 38749103707 scopus 로고    scopus 로고
    • Emotion recognition in spontaneous speech using GMMs
    • In. Pittsburgh, Pennsylvania, 17-19 September
    • Neiberg, D., Elenius, K., & Laskowski, K. (2006). Emotion recognition in spontaneous speech using GMMs. In INTERSPEECH'ICSLP. Pittsburgh, Pennsylvania, 17-19 September 2006 (pp. 809-812).
    • (2006) INTERSPEECH-ICSLP , pp. 809-812
    • Neiberg, D.1    Elenius, K.2    Laskowski, K.3
  • 42
    • 0242721417 scopus 로고    scopus 로고
    • Speech emotion recognition using hidden markov models
    • Nwe, T. L., Foo, S. W., & Silva, L. C. D. (2003). Speech emotion recognition using hidden Markov models. Speech Communication, 41, 603-623.
    • (2003) Speech Communication , vol.41 , pp. 603-623
    • Nwe, T.L.1    Foo, S.W.2    Silva, L.C.D.3
  • 43
    • 0038548330 scopus 로고    scopus 로고
    • The production and recognition of emotions in speech: Features and algorithms
    • Oudeyer, P. Y. (2003). The production and recognition of emotions in speech: Features and algorithms. International Journal ofHuman-Computer Studies, 59, 157-183.
    • (2003) International Journal of Human-Computer Studies , vol.59 , pp. 157-183
    • Oudeyer, P.Y.1
  • 44
    • 33646758219 scopus 로고    scopus 로고
    • Combining acoustic features for improved emotion recognition in mandarin speech
    • In J. Tao, T. Tan, & R. Picard (Eds.), Berlin, Heidelberg, Berlin: Springer
    • Pao, T. L., Chen, Y. T., Yeh, J. H., & Liao, W. Y. (2005). Combining acoustic features for improved emotion recognition in Mandarin speech. In J. Tao, T. Tan, & R. Picard (Eds.), LNCS. ACII, Berlin, Heidelberg (pp. 279-285), Berlin: Springer.
    • (2005) LNCS. ACII , pp. 279-285
    • Pao, T.L.1    Chen, Y.T.2    Yeh, J.H.3    Liao, W.Y.4
  • 45
    • 38049006375 scopus 로고    scopus 로고
    • Feature combination for better differentiating anger from neutral in mandarin emotional speech
    • Berlin: Springer
    • Pao, T. L., Chen, Y. T., Yeh, J. H., Cheng, Y. M., & Chien, C. S. (2007). Feature combination for better differentiating anger from neutral in mandarin emotional speech. In LNCS: Vol. 4738. ACII 2007. Berlin: Springer.
    • (2007) LNCS , vol.4738
    • Pao, T.L.1    Chen, Y.T.2    Yeh, J.H.3    Cheng, Y.M.4    Chien, C.S.5
  • 49
    • 34548794790 scopus 로고    scopus 로고
    • Determination of instants of significant excitation in speech using Hilbert envelope and group delay function
    • DOI 10.1109/LSP.2007.896454
    • Rao, K. S., Prasanna, S. R. M., & Yegnanarayana, B. (2007). Determination of instants of significant excitation in speech using Hilbert envelope and group delay function. IEEE Signal Processing Letters, 14, 762-765. (Pubitemid 47434104)
    • (2007) IEEE Signal Processing Letters , vol.14 , Issue.10 , pp. 762-765
    • Sreenivasa Rao, K.1    Prasanna, S.R.M.2    Yegnanarayana, B.3
  • 51
    • 84864576614 scopus 로고    scopus 로고
    • Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India
    • Reddy, K. S. (2004). Source and system features for speaker recognition. Master's thesis, Department of Computer Science and Engineering, Indian Institute of Technology Madras, Chennai 600 036, India.
    • (2004) Source and System Features for Speaker Recognition
    • Reddy, K.S.1
  • 52
    • 0037384712 scopus 로고    scopus 로고
    • Vocal communication of emotion: A review of research paradigms
    • Scherer, K. R. (2003). Vocal communication of emotion: A review of research paradigms. Speech Communication, 40, 227-256.
    • (2003) Speech Communication , vol.40 , pp. 227-256
    • Scherer, K.R.1
  • 55
    • 70350503956 scopus 로고    scopus 로고
    • Perceived loudness of speech based on the characteristics of glottal excitation source
    • Seshadri, G. P., & Yegnanarayana, B. (2009). Perceived loudness of speech based on the characteristics of glottal excitation source. The Journal of the Acoustical Society of America, 126, 2061-2071.
    • (2009) The Journal of the Acoustical Society of America , vol.126 , pp. 2061-2071
    • Seshadri, G.P.1    Yegnanarayana, B.2
  • 59
    • 0029356550 scopus 로고
    • Usefulness of LPC residue in textindependent speaker verification
    • Thevenaz, P., & Hugli, H. (1995). Usefulness of LPC residue in textindependent speaker verification. Speech Communication, 17, 145-157.
    • (1995) Speech Communication , vol.17 , pp. 145-157
    • Thevenaz, P.1    Hugli, H.2
  • 61
    • 4544247331 scopus 로고    scopus 로고
    • Automatic emotional speech classification
    • New York: IEEE
    • Ververidis, D., Kotropoulos, C., & Pitas, I. (2004). Automatic emotional speech classification. In ICASSP (pp. I593-I596). New York: IEEE.
    • (2004) ICASSP
    • Ververidis, D.1    Kotropoulos, C.2    Pitas, I.3
  • 66
    • 0035989168 scopus 로고    scopus 로고
    • AANN: An alternative to GMM for pattern recognition
    • DOI 10.1016/S0893-6080(02)00019-9, PII S0893608002000199
    • Yegnanarayana, B., & Kishore, S. P. (2002). AANN an alternative to GMM for pattern recognition. Neural Networks, 15, 459-469. (Pubitemid 34518411)
    • (2002) Neural Networks , vol.15 , Issue.3 , pp. 459-469
    • Yegnanarayana, B.1    Kishore, S.P.2
  • 72
    • 38049009485 scopus 로고    scopus 로고
    • Pitch synchronous analysis method and fisher criterion based speaker identification
    • In, Washington DC, USA . Washington: IEEE Computer Society
    • Zeng, Y., Wu, H., & Gao, R. (2007). Pitch synchronous analysis method and Fisher criterion based speaker identification. In Third international conference on natural computation, Washington DC, USA (pp. 691-695). Washington: IEEE Computer Society.
    • (2007) Third International Conference on Natural Computation , pp. 691-695
    • Zeng, Y.1    Wu, H.2    Gao, R.3
  • 73
    • 58849123680 scopus 로고    scopus 로고
    • Emotion recognition in Chinese natural speech by combining prosody and voice quality features
    • In Sun, et al. (Eds.), Berlin: Springer
    • Zhang, S. (2008). Emotion recognition in Chinese natural speech by combining prosody and voice quality features. In Sun, et al. (Eds.), Lecture notes in computer science. Advances in neural networks (pp. 457-464). Berlin: Springer.
    • (2008) Lecture Notes in Computer Science. Advances in Neural Networks , pp. 457-464
    • Zhang, S.1
  • 74
    • 38149033734 scopus 로고    scopus 로고
    • Study on speech emotion recognition system in E-learning
    • In J. Jacko (Ed.), Berlin: Springer
    • Zhu, A., & Luo, Q. (2007). Study on speech emotion recognition system in E-learning. In J. Jacko (Ed.), LNCS. Human computer interaction, Part III, HCII (pp. 544-552). Berlin: Springer.
    • (2007) LNCS. Human Computer Interaction, Part III, HCII , pp. 544-552
    • Zhu, A.1    Luo, Q.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.