메뉴 건너뛰기




Volumn 52, Issue 3, 2013, Pages 1467-1478

Speech emotion recognition approaches in human computer interaction

Author keywords

Human computer interface; Pitch and emotion recognition; Speech emotion

Indexed keywords

ACOUSTIC CHARACTERISTIC; EMOTION RECOGNITION; EMOTIONAL DATABASE; EMOTIONAL SPEECH; EMOTIONAL SPEECH RECOGNITION; HUMAN COMPUTER INTERFACES; SPEECH EMOTION RECOGNITION; SPEECH EMOTIONS;

EID: 84879797312     PISSN: 10184864     EISSN: 15729451     Source Type: Journal    
DOI: 10.1007/s11235-011-9624-z     Document Type: Article
Times cited : (146)

References (50)
  • 2
    • 51849102055 scopus 로고    scopus 로고
    • Automatic recognition of emotions from speech: A review of the literature and recommendations for practical realisation
    • C. Peter R. Beale (eds) LNCS 4868 10.1007/978-3-540-85099-1-7
    • Vogt, T.; Andre, E.; Wagner, J. (2008). Automatic recognition of emotions from speech: a review of the literature and recommendations for practical realisation. In C. Peter & R. Beale (Eds.), LNCS: Vol. 4868. Affect and emotion in HCI (pp. 75-91).
    • (2008) Affect and Emotion in HCI , pp. 75-91
    • Vogt, T.1    Andre, E.2    Wagner, J.3
  • 4
    • 77949760984 scopus 로고    scopus 로고
    • On the classification of emotional biosignals evoked while viewing affective pictures: An integrated data-mining-based approach for healthcare applications
    • 10.1109/TITB.2009.2038481
    • Frantzidis, C. A.; Bratsas, C.; et al. (2010). On the classification of emotional biosignals evoked while viewing affective pictures: an integrated data-mining-based approach for healthcare applications. IEEE Transactions on Information Technology in Biomedicine, 14(2), 309-318.
    • (2010) IEEE Transactions on Information Technology in Biomedicine , vol.14 , Issue.2 , pp. 309-318
    • Frantzidis, C.A.1    Bratsas, C.2
  • 6
    • 72849135549 scopus 로고    scopus 로고
    • Towards an EEG-based emotion recognizer for humanoid robots
    • Toyama, Japan Sept. 27-Oct. 2 University of Karlsruhe (TH), Karlsruhe, Germany
    • Schaaff, K.; Schultz, T. (2009). Towards an EEG-based emotion recognizer for humanoid robots. In The 18th IEEE international symposium on robot and human interactive communication, Toyama, Japan, Sept. 27-Oct. 2 (pp. 719-722). University of Karlsruhe (TH), Karlsruhe, Germany.
    • (2009) The 18th IEEE International Symposium on Robot and Human Interactive Communication , pp. 719-722
    • Schaaff, K.1    Schultz, T.2
  • 11
    • 71049139350 scopus 로고    scopus 로고
    • The Research on Emotion recognition from ECG signal
    • International Conference on Information Technology and Computer Science Kiev July 25-26
    • International Conference on Information Technology and Computer Science (2009). The Research on Emotion recognition from ECG signal. In International conference on information technology and computer science, Kiev, July 25-26
    • (2009) International Conference on Information Technology and Computer Science
  • 12
    • 77953912587 scopus 로고    scopus 로고
    • A new information fusion method for bimodal robotic emotion recognition
    • 10.4304/jcp.3.7.39-47
    • Han, M.-J.; Hsu, J.-H.; Song, K.-T. (2008). A new information fusion method for bimodal robotic emotion recognition. Journal of Computers, 3(7), 39-47.
    • (2008) Journal of Computers , vol.3 , Issue.7 , pp. 39-47
    • Han, M.-J.1    Hsu, J.-H.2    Song, K.-T.3
  • 14
    • 84880628036 scopus 로고    scopus 로고
    • Speech recognition using a wavelet transform to establish fuzzy inference system through subtractive clustering and neural network (ANFIS)
    • Elwakdy, M.; Elsehely, E.; Eltokhy, M.; Elhennawy, A. (2008). Speech recognition using a wavelet transform to establish fuzzy inference system through subtractive clustering and neural network (ANFIS). International Journal of Circuits, Systems and Signal Processing, 4(2), 264-273.
    • (2008) International Journal of Circuits, Systems and Signal Processing , vol.4 , Issue.2 , pp. 264-273
    • Elwakdy, M.1    Elsehely, E.2    Eltokhy, M.3    Elhennawy, A.4
  • 15
    • 84866768695 scopus 로고    scopus 로고
    • Exploring the discrete wavelet transform as a tool for Hindi speech recognition
    • Ranjan, S. (2010). Exploring the discrete wavelet transform as a tool for Hindi speech recognition. International Journal of Computer Theory and Engineering, 2(4), 642-645.
    • (2010) International Journal of Computer Theory and Engineering , vol.2 , Issue.4 , pp. 642-645
    • Ranjan, S.1
  • 17
    • 79958734716 scopus 로고    scopus 로고
    • Context-sensitive multimodal emotion recognition from speech and facial expression using bidirectional lstm modeling
    • Makuhari, Chiba, Japan 26-30 September
    • Wollmer, M.; Metallinou, A.; Eyben, F.; Schuller, B.; Narayanan, S. (2010). Context-sensitive multimodal emotion recognition from speech and facial expression using bidirectional lstm modeling. In International speech communication association, Makuhari, Chiba, Japan, 26-30 September.
    • (2010) International Speech Communication Association
    • Wollmer, M.1    Metallinou, A.2    Eyben, F.3    Schuller, B.4    Narayanan, S.5
  • 20
    • 0037380084 scopus 로고    scopus 로고
    • Emotional speech: Towards a new generation of databases
    • 10.1016/S0167-6393(02)00070-5
    • Douglas-Cowie, E.; Campbell, N.; Cowie, R.; Roach, P. (2003). Emotional speech: towards a new generation of databases. Speech Communication, 40, 33-60.
    • (2003) Speech Communication , vol.40 , pp. 33-60
    • Douglas-Cowie, E.1    Campbell, N.2    Cowie, R.3    Roach, P.4
  • 21
    • 0030283741 scopus 로고    scopus 로고
    • Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition
    • 10.1016/S0167-6393(96)00050-7
    • Hansen, J. H. L. (1996). Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition. Speech Communication, 20(1-2), 151-170.
    • (1996) Speech Communication , vol.20 , Issue.1-2 , pp. 151-170
    • Hansen, J.H.L.1
  • 22
    • 65249116503 scopus 로고    scopus 로고
    • Analysis of emotionally salient aspects of fundamental frequency for emotion detection
    • 10.1109/TASL.2008.2009578
    • Busso, C.; Lee, S.; Narayanan, S. (2009). Analysis of emotionally salient aspects of fundamental frequency for emotion detection. IEEE Transactions on Audio, Speech, and Language Processing, 17(4), 582-596.
    • (2009) IEEE Transactions on Audio, Speech, and Language Processing , vol.17 , Issue.4 , pp. 582-596
    • Busso, C.1    Lee, S.2    Narayanan, S.3
  • 25
    • 50049092345 scopus 로고    scopus 로고
    • Fast and accurate sequential floating forward feature selection with the Bayes classifier applied to speech emotion recognition
    • 10.1016/j.sigpro.2008.07.001
    • Ververidis, D.; Kotropoulos, C. (2008). Fast and accurate sequential floating forward feature selection with the Bayes classifier applied to speech emotion recognition. Signal Processing, 88(12), 2956-2970.
    • (2008) Signal Processing , vol.88 , Issue.12 , pp. 2956-2970
    • Ververidis, D.1    Kotropoulos, C.2
  • 26
    • 0037697289 scopus 로고    scopus 로고
    • A spatio-temporal speech enhancement scheme for robust speech recognition in noisy environments
    • 10.1016/S0167-6393(03)00010-4
    • Visser, E.; Otsuka, M.; Lee, T.-W. (2003). A spatio-temporal speech enhancement scheme for robust speech recognition in noisy environments. Speech Communication, 41, 393-407.
    • (2003) Speech Communication , vol.41 , pp. 393-407
    • Visser, E.1    Otsuka, M.2    Lee, T.-W.3
  • 31
    • 68349121449 scopus 로고    scopus 로고
    • Emotion recognition using multilayer perceptron and generalized feed forward neural network
    • Khanchandani, K. B.; Hussain, M. A. (2009). Emotion recognition using multilayer perceptron and generalized feed forward neural network. IEEE Journal of Scientific and Industrial Research, 68, 367-371.
    • (2009) IEEE Journal of Scientific and Industrial Research , vol.68 , pp. 367-371
    • Khanchandani, K.B.1    Hussain, M.A.2
  • 32
    • 77952636818 scopus 로고    scopus 로고
    • Classification of complex information: Inference of co-occurring affective states from their expressions in speech
    • 10.1109/TPAMI.2009.107
    • Sobol-Shikler, T.; Robinson, P. (2010). Classification of complex information: inference of co-occurring affective states from their expressions in speech. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(7), 1284-1297.
    • (2010) IEEE Transactions on Pattern Analysis and Machine Intelligence , vol.32 , Issue.7 , pp. 1284-1297
    • Sobol-Shikler, T.1    Robinson, P.2
  • 33
    • 58249124688 scopus 로고    scopus 로고
    • Emotional intelligence, not music training, predicts recognition of emotional speech prosody
    • 10.1037/a0014080 Copyright 2008 by the American Psychological Association
    • Trimmer, C. G.; Cuddy, L. L. (2008). Emotional intelligence, not music training, predicts recognition of emotional speech prosody. Emotion, 8(6), 838-849. Copyright 2008 by the American Psychological Association.
    • (2008) Emotion , vol.8 , Issue.6 , pp. 838-849
    • Trimmer, C.G.1    Cuddy, L.L.2
  • 37
    • 57849145227 scopus 로고    scopus 로고
    • Automatic recognition of spontaneous emotions in speech using acoustic and lexical features
    • LNCS 5237
    • Truong, K. P.; Raaijmakers, S. (2008). Automatic recognition of spontaneous emotions in speech using acoustic and lexical features. In LNCS: Vol. 5237. MLMI 2008 (pp. 161-172).
    • (2008) MLMI 2008 , pp. 161-172
    • Truong, K.P.1    Raaijmakers, S.2
  • 38
    • 4544316885 scopus 로고    scopus 로고
    • Speech emotion recognition combining acoustic features and linguistic information in a hybrid support vector machine - Belief network architecture
    • Quebec, Canada 17-21 May
    • Schuller, B.; Rigoll, G.; Lang, M. (2004). Speech emotion recognition combining acoustic features and linguistic information in a hybrid support vector machine - belief network architecture. In IEEE international conference on acoustics, speech, and signal processing, Quebec, Canada, 17-21 May.
    • (2004) IEEE International Conference on Acoustics, Speech, and Signal Processing
    • Schuller, B.1    Rigoll, G.2    Lang, M.3
  • 39
    • 54049132987 scopus 로고    scopus 로고
    • Combining speech recognition and acoustic word emotion models for robust text-independent emotion recognition
    • Hannover, Germany 23-26 June
    • Schuller, B.; Vlasenko, B.; Arsic, D.; Rigoll, G.; Wendemuth, A. (2008). Combining speech recognition and acoustic word emotion models for robust text-independent emotion recognition. In IEEE international conference on multimedia & expo, Hannover, Germany, 23-26 June.
    • (2008) IEEE International Conference on Multimedia & Expo
    • Schuller, B.1    Vlasenko, B.2    Arsic, D.3    Rigoll, G.4    Wendemuth, A.5
  • 42
    • 67650159383 scopus 로고    scopus 로고
    • Improved emotion recognition with a novel speaker-independent feature
    • 10.1109/TMECH.2008.2008644
    • Kim, E. H.; Hyun, K. H.; Kim, S. H.; Kwak, Y. K. (2009). Improved emotion recognition with a novel speaker-independent feature. IEEE/ASME Transactions on Mechatronics, 14(3), 317-325.
    • (2009) IEEE/ASME Transactions on Mechatronics , vol.14 , Issue.3 , pp. 317-325
    • Kim, E.H.1    Hyun, K.H.2    Kim, S.H.3    Kwak, Y.K.4
  • 43
    • 77956721304 scopus 로고    scopus 로고
    • Combining long short-term memory and dynamic Bayesian networks for incremental emotion-sensitive artificial listening
    • 10.1109/JSTSP.2010.2057200
    • Wöllmer, M.; Schuller, B.; Eyben, F.; Rigoll, G. (2010). Combining long short-term memory and dynamic Bayesian networks for incremental emotion-sensitive artificial listening. IEEE Journal of Selected Topics in Signal Processing, 4(5), 867-881.
    • (2010) IEEE Journal of Selected Topics in Signal Processing , vol.4 , Issue.5 , pp. 867-881
    • Wöllmer, M.1    Schuller, B.2    Eyben, F.3    Rigoll, G.4
  • 46
    • 40349093030 scopus 로고    scopus 로고
    • How aging affects the recognition of emotional speech
    • 10.1016/j.bandl.2007.03.002
    • Paulmann, S.; Pell, M. D.; Kotz, S. A. (2008). How aging affects the recognition of emotional speech. Brain and Language, 104, 262-269.
    • (2008) Brain and Language , vol.104 , pp. 262-269
    • Paulmann, S.1    Pell, M.D.2    Kotz, S.A.3
  • 47
    • 37349079113 scopus 로고    scopus 로고
    • Critical analysis of the impact of glottal features in the classification of clinical depression in speech
    • 10.1109/TBME.2007.900562
    • Moore, E.; II, Clements, M. A.; Peifer, J. W.; Weisser, L. (2008). Critical analysis of the impact of glottal features in the classification of clinical depression in speech. IEEE Transactions on Biomedical Engineering, 55(1), 96-107.
    • (2008) IEEE Transactions on Biomedical Engineering , vol.55 , Issue.1 , pp. 96-107
    • Moore, I.I.E.1    Clements, M.A.2    Peifer, J.W.3    Weisser, L.4
  • 48
    • 84879793485 scopus 로고    scopus 로고
    • Speech emotion recognition for affective human-robot interaction
    • St. Petersburg 25-29 June
    • Jang, K.-D.; Kwon, O.-W. (2006). Speech emotion recognition for affective human-robot interaction. In SPECOM'2006, St. Petersburg, 25-29 June (pp. 419-422).
    • (2006) SPECOM'2006 , pp. 419-422
    • Jang, K.-D.1    Kwon, O.-W.2
  • 49
    • 70350300961 scopus 로고    scopus 로고
    • Feature vector classification based speech emotion recognition for service robots
    • 10.1109/TCE.2009.5278031
    • Park, J.-S.; Kim, J.-H.; Oh, Y.-H. (2009). Feature vector classification based speech emotion recognition for service robots. IEEE Transactions on Consumer Electronics, 55(3), 1590-1596.
    • (2009) IEEE Transactions on Consumer Electronics , vol.55 , Issue.3 , pp. 1590-1596
    • Park, J.-S.1    Kim, J.-H.2    Oh, Y.-H.3
  • 50
    • 47649091996 scopus 로고    scopus 로고
    • Recognizing human emotional state from audiovisual signals
    • 10.1109/TMM.2008.921734
    • Wang, Y.; Guan, L. (2008). Recognizing human emotional state from audiovisual signals. IEEE Transactions on Multimedia, 10(4), 659-668.
    • (2008) IEEE Transactions on Multimedia , vol.10 , Issue.4 , pp. 659-668
    • Wang, Y.1    Guan, L.2


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.