메뉴 건너뛰기




Volumn 2016-January, Issue , 2016, Pages 2196-2202

Discriminatively trained recurrent neural networks for continuous dimensional emotion recognition from audio

Author keywords

[No Author keywords available]

Indexed keywords

ARTIFICIAL INTELLIGENCE; COST FUNCTIONS; ERRORS; MEAN SQUARE ERROR; NEURAL NETWORKS; SPEECH RECOGNITION;

EID: 85006110321     PISSN: 10450823     EISSN: None     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (60)

References (27)
  • 1
    • 85014296970 scopus 로고    scopus 로고
    • A neural network model for the prediction of musical emotions
    • S. Nefti-Meziani and J. G. Grey, editors, IET Publisher, London, UK
    • Eduardo Coutinho and Angelo Cangelosi. A neural network model for the prediction of musical emotions. In S. Nefti-Meziani and J. G. Grey, editors, Advances in Cognitive Systems, pages 331-368. IET Publisher, London, UK, 2010.
    • (2010) Advances in Cognitive Systems , pp. 331-368
    • Coutinho, E.1    Cangelosi, A.2
  • 4
    • 0034293152 scopus 로고    scopus 로고
    • Learning to forget: Continual prediction with LSTM
    • Felix A. Gers, Jürgen Schmidhuber, and Fred Cummins. Learning to forget: Continual prediction with LSTM. Neural Computation, 12 (10): 2451-2471, 2000.
    • (2000) Neural Computation , vol.12 , Issue.10 , pp. 2451-2471
    • Gers, F.A.1    Schmidhuber, J.2    Cummins, F.3
  • 5
    • 84910060363 scopus 로고    scopus 로고
    • Speech emotion recognition using deep neural network and extreme learning machine
    • Singapore, September. ISCA
    • Kun Han, Dong Yu, and Ivan Tashev. Speech Emotion Recognition Using Deep Neural Network and Extreme Learning Machine. In Proc. of INTERSPEECH, pages 223-227, Singapore, September 2014. ISCA.
    • (2014) Proc. of INTERSPEECH , pp. 223-227
    • Han, K.1    Yu, D.2    Tashev, I.3
  • 6
    • 84960882502 scopus 로고    scopus 로고
    • Multimodal affective dimension prediction using deep bidirectional long short-term memory recurrent neural networks
    • Brisbane, Australia, October. ACM
    • Lang He, Dongmei Jiang, Le Yang, Ercheng Pei, Peng Wu, and Hichem Sahli. Multimodal Affective Dimension Prediction Using Deep Bidirectional Long Short-Term Memory Recurrent Neural Networks. In Proc. of AVEC, pages 73-80, Brisbane, Australia, October 2015. ACM.
    • (2015) Proc. of AVEC , pp. 73-80
    • He, L.1    Jiang, D.2    Yang, L.3    Pei, E.4    Wu, P.5    Sahli, H.6
  • 7
    • 84941334839 scopus 로고    scopus 로고
    • Joint opti-mization of masks and deep recurrent neural networks for monaural source separation
    • December
    • Po-Sen Huang, Minje Kim, Mark Hasegawa-Johnson, and Paris Smaragdis. Joint Opti-mization of Masks and Deep Recurrent Neural Networks for Monaural Source Separation. IEEE Transactions on Audio, Speech, and Language Processing, 23 (12): 2136-2147, December 2015.
    • (2015) IEEE Transactions on Audio, Speech, and Language Processing , vol.23 , Issue.12 , pp. 2136-2147
    • Huang, P.-S.1    Kim, M.2    Mark, H.-J.3    Smaragdis, P.4
  • 8
    • 84893298548 scopus 로고    scopus 로고
    • A comparison of evaluation measures for emotion recognition in dimensional space
    • Geneva, Switzerland, September. IEEE
    • Robert Jenke, Angelika Peer, and Martin Buss. A Comparison of Evaluation Measures for Emotion Recognition in Dimensional Space. In Proc. of ACII, pages 822-826, Geneva, Switzerland, September 2013. IEEE.
    • (2013) Proc. of ACII , pp. 822-826
    • Jenke, R.1    Peer, A.2    Buss, M.3
  • 9
    • 84960847562 scopus 로고    scopus 로고
    • Ensemble methods for continuous affect recognition: Multimodality, temporality, and challenges
    • Brisbane, Australia, October. ACM
    • Markus Kächele, Patrick Thiam, Günther Palm, Friedhelm Schwenker, and Martin Schels. Ensemble methods for continuous affect recognition: Multimodality, temporality, and challenges. In Proc. of AVEC (held in conjunction with ACM MM), pages 9-16, Brisbane, Australia, October 2015. ACM.
    • (2015) Proc. of AVEC (Held in Conjunction with ACM MM) , pp. 9-16
    • Kächele, M.1    Thiam, P.2    Palm, G.3    Schwenker, F.4    Schels, M.5
  • 10
    • 85006140664 scopus 로고    scopus 로고
    • Scalable minimum Bayes risk training of deep neural network acoustic models using distributed Hessian-free optimization
    • Kyoto, Japan
    • Brian Kingsbury, Tara N. Sainath, and Hagen Soltau. Scalable minimum Bayes risk training of deep neural network acoustic models using distributed Hessian-free optimization. In Proc. of ICASSP, Kyoto, Japan, 2012.
    • (2012) Proc. of ICASSP
    • Kingsbury, B.1    Sainath, T.N.2    Soltau, H.3
  • 11
    • 84893307972 scopus 로고    scopus 로고
    • Hybrid deep neural network-hidden markov model (dnn-hmm) based speech emotion recognition
    • Geneva, Switzerland, September. IEEE
    • Longfei Li, Yong Zhao, Dongmei Jiang, Yanning Zhang, Fengna Wang, I. Gonzalez, E. Valentin, and H. Sahli. Hybrid Deep Neural Network-Hidden Markov Model (DNN-HMM) Based Speech Emotion Recognition. In Proc. of ACII, pages 312-317, Geneva, Switzerland, September 2013. IEEE.
    • (2013) Proc. of ACII , pp. 312-317
    • Li, L.1    Zhao, Y.2    Jiang, D.3    Zhang, Y.4    Wang, F.5    Gonzalez, I.6    Valentin, E.7    Sahli, H.8
  • 12
    • 0024521543 scopus 로고
    • A concordance correlation coefficient to evaluate reproducibility
    • March
    • Lawrence I. Lin. A concordance correlation coefficient to evaluate reproducibility. Biometrics, 45 (1): 255-268, March 1989.
    • (1989) Biometrics , vol.45 , Issue.1 , pp. 255-268
    • Lin, L.I.1
  • 13
    • 84930944930 scopus 로고    scopus 로고
    • Correcting time-continuous emotional labels by modeling the reaction lag of evaluators
    • April-June
    • Soroosh Mariooryad and Carlos Busso. Correcting time-continuous emotional labels by modeling the reaction lag of evaluators. IEEE Transactions on Affective Computing, 6 (2): 97-108, April-June 2015.
    • (2015) IEEE Transactions on Affective Computing , vol.6 , Issue.2 , pp. 97-108
    • Mariooryad, S.1    Busso, C.2
  • 14
    • 84886390329 scopus 로고    scopus 로고
    • Tracking continuous emotional trends of participants during affective dyadic interactions using body language and speech information
    • February
    • Angeliki Metallinou, Athanasios Katsamanis, and Shrikanth Narayanan. Tracking continuous emotional trends of participants during affective dyadic interactions using body language and speech information. Image and Vision Computing, 31 (2): 137-152, February 2013.
    • (2013) Image and Vision Computing , vol.31 , Issue.2 , pp. 137-152
    • Metallinou, A.1    Katsamanis, A.2    Narayanan, S.3
  • 15
    • 84881512394 scopus 로고    scopus 로고
    • Introducing the RECOLA Multimodal Corpus of Remote Collaborative and Affective Interactions
    • Shanghai, China, April
    • Fabien Ringeval, Andreas Sonderegger, Jürgen Sauer, and Denis Lalanne. Introducing the RECOLA Multimodal Corpus of Remote Collaborative and Affective Interactions. In Proc. of EmoSPACE (held in conjunction with ACM FG), Shanghai, China, April 2013. 8 pages.
    • (2013) Proc. of EmoSPACE (Held in Conjunction with ACM FG)
    • Ringeval, F.1    Sonderegger, A.2    Sauer, J.3    Lalanne, D.4
  • 19
    • 84872700353 scopus 로고    scopus 로고
    • Modeling musical emotion dynamics with conditional random fields
    • Miami, FL, USA
    • Erik M. Schmidt and Youngmoo E. Kim. Modeling musical emotion dynamics with conditional random fields. In Proc. of ISMIR, pages 777-782, Miami, FL, USA, 2011.
    • (2011) Proc. of ISMIR , pp. 777-782
    • Schmidt, E.M.1    Kim, Y.E.2
  • 20
    • 84873420121 scopus 로고    scopus 로고
    • Feature learning in dynamic environments: Modeling the acoustic structure of musical emotion
    • Miami, FL, USA
    • Erik M. Schmidt, Jeffrey Scott, and Youngmoo E. Kim. Feature learning in dynamic environments: Modeling the acoustic structure of musical emotion. In Proc. of ISMIR, pages 325-330, Miami, FL, USA, 2012.
    • (2012) Proc. of ISMIR , pp. 325-330
    • Schmidt, E.M.1    Scott, J.2    Kim, Y.E.3
  • 21
    • 84870223036 scopus 로고    scopus 로고
    • AVEC 2012-the continuous audio/visual emotion challenge
    • Louis-Philippe Morency, Dan Bohus, Hamid K. Aghajan, Justine Cassell, Anton Nijholt, and Julien Epps, editors, Santa Monica, CA, October. ACM
    • Björn Schuller, Michel Valstar, Florian Eyben, Roddy Cowie, and Maja Pantic. AVEC 2012-The Continuous Audio/Visual Emotion Challenge. In Louis-Philippe Morency, Dan Bohus, Hamid K. Aghajan, Justine Cassell, Anton Nijholt, and Julien Epps, editors, Proc. of ICMI, pages 449-456, Santa Monica, CA, October 2012. ACM.
    • (2012) Proc. of ICMI , pp. 449-456
    • Schuller, B.1    Valstar, M.2    Eyben, F.3    Cowie, R.4    Pantic, M.5
  • 23
    • 84906274730 scopus 로고    scopus 로고
    • Sequence-discriminative training of deep neural networks
    • Lyon, France. ISCA
    • Karek Veselý, Arnab Ghoshal, Lukás Burget, and Daniel Povey. Sequence-discriminative training of deep neural networks. In Proc. of INTERSPEECH, pages 2345-2349, Lyon, France, 2013. ISCA.
    • (2013) Proc. of INTERSPEECH , pp. 2345-2349
    • Veselý, K.1    Ghoshal, A.2    Burget, L.3    Povey, D.4
  • 24
    • 84945462625 scopus 로고    scopus 로고
    • Relevance units machine based dimensional and continuous speech emotion prediction
    • November
    • Fengna Wang, Hichem Sahli, Junbin Gao, Dongmei Jiang, and Werner Verhelst. Relevance units machine based dimensional and continuous speech emotion prediction. Multimedia Tools and Applications, 74 (22): 9983-10000, November 2015.
    • (2015) Multimedia Tools and Applications , vol.74 , Issue.22 , pp. 9983-10000
    • Wang, F.1    Sahli, H.2    Gao, J.3    Jiang, D.4    Verhelst, W.5
  • 25
    • 85006140676 scopus 로고    scopus 로고
    • The TUM approach to the MediaEval music emotion task using generic affective audio features
    • Martha Larson, Xavier Anguera, Timo Reuter, Gareth J. F. Jones, Bogdan Ionescu, Markus Schedl, Tomas Piatrik, Claudia Hauff, and Mohammad Soleymani, editors, Barcelona, Spain, October. CEUR
    • Felix Weninger, Florian Eyben, and Björn Schuller. The TUM approach to the MediaEval music emotion task using generic affective audio features. In Martha Larson, Xavier Anguera, Timo Reuter, Gareth J. F. Jones, Bogdan Ionescu, Markus Schedl, Tomas Piatrik, Claudia Hauff, and Mohammad Soleymani, editors, Proc. of MediaEval, Barcelona, Spain, October 2013. CEUR.
    • (2013) Proc. of MediaEval
    • Weninger, F.1    Eyben, F.2    Schuller, B.3
  • 26
    • 84941334311 scopus 로고    scopus 로고
    • Discriminatively trained recurrent neural networks for single-channel speech separation
    • Atlanta, GA, USA. IEEE
    • Felix Weninger, John R. Hershey, Jonathan Le Roux, and Björn Schuller. Discriminatively trained recurrent neural networks for single-channel speech separation. In Proc. of GlobalSIP, pages 740-744, Atlanta, GA, USA, 2014. IEEE.
    • (2014) Proc. of GlobalSIP , pp. 740-744
    • Weninger, F.1    Hershey, J.R.2    Le Roux, J.3    Schuller, B.4
  • 27
    • 84930639546 scopus 로고    scopus 로고
    • Introducing CURRENNT-the Munich open-source CUDA RecurREnt neural network toolkit
    • Felix Weninger, Johannes Bergmann, and Björn Schuller. Introducing CURRENNT-the Munich open-source CUDA RecurREnt Neural Network Toolkit. Journal of Machine Learning Research, 16: 547-551, 2015.
    • (2015) Journal of Machine Learning Research , vol.16 , pp. 547-551
    • Weninger, F.1    Bergmann, J.2    Schuller, B.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.