-
2
-
-
0030355346
-
Characterizing audiovisual information during speech
-
Philadelphia, PA, Oct
-
E. Vatikiotis-Bateson, K. Munhall, Y. Kasahara, F. Garcia, and H. Yehia, "Characterizing audiovisual information during speech," in Proc. 4th Int. Conf. Spoken Lang. Process. (ICSLP 96), Philadelphia, PA, Oct. 1996, vol. 3, pp. 1485-1488.
-
(1996)
Proc. 4th Int. Conf. Spoken Lang. Process. (ICSLP 96)
, vol.3
, pp. 1485-1488
-
-
Vatikiotis-Bateson, E.1
Munhall, K.2
Kasahara, Y.3
Garcia, F.4
Yehia, H.5
-
3
-
-
85029641276
-
Animated conversation: Rule-based generation of facial expression gesture and spoken intonation for multiple conversational agents
-
Orlando, FL
-
J. Cassell, C. Pelachaud, N. Badler, M. Steedman, B. Achorn, T. Bechet, B. Douville, S. Prevost, and M. Stone, "Animated conversation: Rule-based generation of facial expression gesture and spoken intonation for multiple conversational agents," in Comput. Graphics (Proc. ACM SIGGRAPH'94), Orlando, FL, 1994, pp. 413-420.
-
(1994)
Comput. Graphics (Proc. ACM SIGGRAPH'94)
, pp. 413-420
-
-
Cassell, J.1
Pelachaud, C.2
Badler, N.3
Steedman, M.4
Achorn, B.5
Bechet, T.6
Douville, B.7
Prevost, S.8
Stone, M.9
-
4
-
-
84960898014
-
Multimodal signal analysis of prosody and hand motion: Temporal correlation of speech and gestures
-
Tolouse, France, Sep
-
L. Valbonesi, R. Ansari, D. McNeill, F. Quek, S. Duncan, K. McCullough, and R. Bryll, "Multimodal signal analysis of prosody and hand motion: Temporal correlation of speech and gestures," in Proc. Eur. Signal Process. Conf. (EUSIPCO 02), Tolouse, France, Sep. 2002, pp. 75-78.
-
(2002)
Proc. Eur. Signal Process. Conf. (EUSIPCO 02)
, pp. 75-78
-
-
Valbonesi, L.1
Ansari, R.2
McNeill, D.3
Quek, F.4
Duncan, S.5
McCullough, K.6
Bryll, R.7
-
5
-
-
0027588084
-
Facial expression and emotion
-
Apr
-
P. Ekman, "Facial expression and emotion," Amer. Psychol., vol. 48, no. 4, pp. 384-392, Apr. 1993.
-
(1993)
Amer. Psychol
, vol.48
, Issue.4
, pp. 384-392
-
-
Ekman, P.1
-
7
-
-
42949167982
-
Rigid head motionin expressive speech animation: Analysis and synthesis
-
Mar
-
C. Busso, Z. Deng, M. Grimm, U. Neumann, and S. Narayanan, "Rigid head motionin expressive speech animation: Analysis and synthesis," IEEE Trans. Audio, Speech, Lang. Process., vol. 15, no. 3, pp. 1075-1086, Mar. 2007.
-
(2007)
IEEE Trans. Audio, Speech, Lang. Process
, vol.15
, Issue.3
, pp. 1075-1086
-
-
Busso, C.1
Deng, Z.2
Grimm, M.3
Neumann, U.4
Narayanan, S.5
-
8
-
-
0037382510
-
Describing the emotional states that are expressed in speech
-
Apr
-
R. Cowie and R. Cornelius, "Describing the emotional states that are expressed in speech," Speech Commun., vol. 40, no. 1-2, pp. 5-32, Apr. 2003.
-
(2003)
Speech Commun
, vol.40
, Issue.1-2
, pp. 5-32
-
-
Cowie, R.1
Cornelius, R.2
-
9
-
-
0037384712
-
Vocal communication of emotion: A review of research paradigms
-
Apr
-
K. Scherer, "Vocal communication of emotion: A review of research paradigms," Speech Commun., vol. 40, no. 1-2, pp. 227-256, Apr. 2003.
-
(2003)
Speech Commun
, vol.40
, Issue.1-2
, pp. 227-256
-
-
Scherer, K.1
-
10
-
-
33745190613
-
An acoustic study of emotions expressed in speech
-
Jeju Island, Korea
-
S. Yildirim, M. Bulut, C. Lee, A. Kazemzadeh, C. Busso, Z. Deng, S. Lee, and S. Narayanan, "An acoustic study of emotions expressed in speech," in Proc. 8th Int. Conf. Spoken Lang. Process. (ICSLP 04), Jeju Island, Korea, 2004, pp. 2193-2196.
-
(2004)
Proc. 8th Int. Conf. Spoken Lang. Process. (ICSLP 04)
, pp. 2193-2196
-
-
Yildirim, S.1
Bulut, M.2
Lee, C.3
Kazemzadeh, A.4
Busso, C.5
Deng, Z.6
Lee, S.7
Narayanan, S.8
-
11
-
-
0002142055
-
About Brows: Emotional and Conversational Signals
-
M. von Cranach, K. Foppa, W. Lepenies, and D. Ploog, Eds. New York: Cambridge Univ. Press
-
P. Ekman, "About Brows: Emotional and Conversational Signals," in Human Ethology: Claims and Limits of a New Discipline, M. von Cranach, K. Foppa, W. Lepenies, and D. Ploog, Eds. New York: Cambridge Univ. Press, 1979, pp. 169-202.
-
(1979)
Human Ethology: Claims and Limits of a New Discipline
, pp. 169-202
-
-
Ekman, P.1
-
12
-
-
0032178592
-
Quantitative association of vocal-tract and facial behavior
-
H. Yehia, P. Rubin, and E. Vatikiotis-Bateson, "Quantitative association of vocal-tract and facial behavior," Speech Commun., vol. 26, no. 1-2, pp. 23-43, 1998.
-
(1998)
Speech Commun
, vol.26
, Issue.1-2
, pp. 23-43
-
-
Yehia, H.1
Rubin, P.2
Vatikiotis-Bateson, E.3
-
13
-
-
0036656895
-
Linking facial animation, head motion and speech acoustics
-
Jul
-
H. Yehia, T. Kuratate, and E. Vatikiotis-Bateson, "Linking facial animation, head motion and speech acoustics," J. Phonetics, vol. 30, no. 3, pp. 555-568, Jul. 2002.
-
(2002)
J. Phonetics
, vol.30
, Issue.3
, pp. 555-568
-
-
Yehia, H.1
Kuratate, T.2
Vatikiotis-Bateson, E.3
-
14
-
-
0030369502
-
About the relationship between eyebrow movements and F0 variations
-
Philadelphia, PA, Oct
-
C. Cavé, I. Guaïtella, R. Bertrand, S. Santi, F. Harlay, and R. Espesser, "About the relationship between eyebrow movements and F0 variations," in Int. Conf. Spoken Lang. Process. (ICSLP), Philadelphia, PA, Oct. 1996, vol. 4, pp. 2175-2178.
-
(1996)
Int. Conf. Spoken Lang. Process. (ICSLP)
, vol.4
, pp. 2175-2178
-
-
Cavé, C.1
Guaïtella, I.2
Bertrand, R.3
Santi, S.4
Harlay, F.5
Espesser, R.6
-
15
-
-
14944351245
-
Analysis of emotion recognition using facial expressions, speech and multimodal information
-
State College, PA
-
C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. Lee, A. Kazemzadeh, S. Lee, U. Neumann, and S. Narayanan, "Analysis of emotion recognition using facial expressions, speech and multimodal information," in Proc. 6th Int. Conf. Multimodal Interfaces ICMI 04., State College, PA, 2004, pp. 205-211.
-
(2004)
Proc. 6th Int. Conf. Multimodal Interfaces ICMI 04
, pp. 205-211
-
-
Busso, C.1
Deng, Z.2
Yildirim, S.3
Bulut, M.4
Lee, C.5
Kazemzadeh, A.6
Lee, S.7
Neumann, U.8
Narayanan, S.9
-
16
-
-
0032634966
-
Embodiment in conversational interfaces: Rea
-
Pittsburgh, PA, May
-
J. Cassell, T. Bickmore, M. Billinghurst, L. Campbell, K. Chang, H. Vilhjalmsson, and H. Yan, "Embodiment in conversational interfaces: Rea," in Proc. Int. Conf. Human Factors in Comput. Syst. (CHI 99), Pittsburgh, PA, May 1999, pp. 520-527.
-
(1999)
Proc. Int. Conf. Human Factors in Comput. Syst. (CHI 99)
, pp. 520-527
-
-
Cassell, J.1
Bickmore, T.2
Billinghurst, M.3
Campbell, L.4
Chang, K.5
Vilhjalmsson, H.6
Yan, H.7
-
17
-
-
9444285497
-
Towards symmetric multi-modality: Fusion and fission of speech, gesture, and facial expression
-
W. Wahlster, A. Günter, R. Kruse, and B. Neumann, Eds, Berlin, Germany
-
W. Wahlster, A. Günter, R. Kruse, and B. Neumann, Eds., "Towards symmetric multi-modality: Fusion and fission of speech, gesture, and facial expression," in Proc. 26th German Conf. Artif. Intell., Berlin, Germany, 2003, pp. 1-18.
-
(2003)
Proc. 26th German Conf. Artif. Intell
, pp. 1-18
-
-
-
18
-
-
27144506606
-
Natural head motion synthesis driven by acoustic prosodic features
-
Jul
-
C. Busso, Z. Deng, U. Neumann, and S. Narayanan, "Natural head motion synthesis driven by acoustic prosodic features," Comput. Animation Virtual Worlds, vol. 16, no. 3-4, pp. 283-290, Jul. 2005.
-
(2005)
Comput. Animation Virtual Worlds
, vol.16
, Issue.3-4
, pp. 283-290
-
-
Busso, C.1
Deng, Z.2
Neumann, U.3
Narayanan, S.4
-
19
-
-
78650465043
-
Visual prosody: Facialmovements accompanying speech
-
Washington, DC, May
-
H. P. Graf, E. Cosatto, V. Strom, and F. J. Huang, "Visual prosody: Facialmovements accompanying speech," in Proc. IEEE Int. Conf. Autom. Faces and Gesture Recognition, Washington, DC, May 2002, pp. 396-401.
-
(2002)
Proc. IEEE Int. Conf. Autom. Faces and Gesture Recognition
, pp. 396-401
-
-
Graf, H.P.1
Cosatto, E.2
Strom, V.3
Huang, F.J.4
-
20
-
-
22144492019
-
Expressive audio-visual speech
-
Jul
-
E. Bevacqua and C. Pelachaud, "Expressive audio-visual speech," Comput. Animation Virtual Worlds, vol. 15, no. 3-4, pp. 297-304, Jul. 2004.
-
(2004)
Comput. Animation Virtual Worlds
, vol.15
, Issue.3-4
, pp. 297-304
-
-
Bevacqua, E.1
Pelachaud, C.2
-
21
-
-
13144278330
-
Speech-driven facial animation with realistic dynamics
-
Feb
-
R. Gutierrez-Osuna, P. Kakumanu, A. Esposito, O. Garcia, A. Bojorquez, J. Castillo, and I. Rudomin, "Speech-driven facial animation with realistic dynamics," IEEE Trans. Multimedia, vol. 7, no. 1, pp. 33-42, Feb. 2005.
-
(2005)
IEEE Trans. Multimedia
, vol.7
, Issue.1
, pp. 33-42
-
-
Gutierrez-Osuna, R.1
Kakumanu, P.2
Esposito, A.3
Garcia, O.4
Bojorquez, A.5
Castillo, J.6
Rudomin, I.7
-
22
-
-
33745191649
-
An articulatory studyof emotional speech production
-
Lisbon, Portugal, Sep
-
S. Lee, S. Yildirim, A. Kazemzadeh, and S. Narayanan, "An articulatory studyof emotional speech production," in 9th Eur. Conf. Speech Commun. Technol. (Interspeech'05-Eurospeech), Lisbon, Portugal, Sep. 2005, pp. 497-500.
-
(2005)
9th Eur. Conf. Speech Commun. Technol. (Interspeech'05-Eurospeech)
, pp. 497-500
-
-
Lee, S.1
Yildirim, S.2
Kazemzadeh, A.3
Narayanan, S.4
-
23
-
-
21844443583
-
Audiovisual representation of prosody inexpressive speech communication
-
Jul
-
B. Granström and D. House, "Audiovisual representation of prosody inexpressive speech communication," Speech Commun., vol. 46, no. 3-4, pp. 473-484, Jul. 2005.
-
(2005)
Speech Commun
, vol.46
, Issue.3-4
, pp. 473-484
-
-
Granström, B.1
House, D.2
-
24
-
-
0036874551
-
On the relationship between face movements, tongue movements, and speech acoustics
-
J. Jiang, A. Alwan, P. Keating, B. Chaney, E. Auer, Jr, and L. Bernstein, "On the relationship between face movements, tongue movements, and speech acoustics," EURASIP J. Appl. Signal Process., vol. 11, pp. 1174-1188, 2002.
-
(2002)
EURASIP J. Appl. Signal Process
, vol.11
, pp. 1174-1188
-
-
Jiang, J.1
Alwan, A.2
Keating, P.3
Chaney, B.4
Auer Jr, E.5
Bernstein, L.6
-
25
-
-
0012725678
-
Estimation of speech acoustics from visual speech features: A comparison of linear and nonlinear models
-
Santa Cruz, CA, Aug
-
J. P. Barker and F. Berthommier, "Estimation of speech acoustics from visual speech features: A comparison of linear and nonlinear models," in Conf. Audio-Visual Speech Process. (AVSP 99), Santa Cruz, CA, Aug. 1999, pp. 112-117.
-
(1999)
Conf. Audio-Visual Speech Process. (AVSP 99)
, pp. 112-117
-
-
Barker, J.P.1
Berthommier, F.2
-
26
-
-
33846592328
-
Audio-visual affect recognition
-
Feb
-
Z. Zeng, J. Tu,M. Liu, T. Huang, B. Pianfetti, D. Roth, and S. Levinson, "Audio-visual affect recognition," IEEE Trans. Multimedia, vol. 9, no. 2, pp. 424-428, Feb. 2007.
-
(2007)
IEEE Trans. Multimedia
, vol.9
, Issue.2
, pp. 424-428
-
-
Zeng, Z.1
Tu, J.2
Liu, M.3
Huang, T.4
Pianfetti, B.5
Roth, D.6
Levinson, S.7
-
27
-
-
4544290191
-
Recent advancesin the automatic recognition of audiovisual speech
-
Sep
-
G. Potamianos, C. Neti, G. Gravier,A. Garg, and A. Senior, "Recent advancesin the automatic recognition of audiovisual speech," Proc. IEEE, vol. 91, no. 9, pp. 1306-1326, Sep. 2003.
-
(2003)
Proc. IEEE
, vol.91
, Issue.9
, pp. 1306-1326
-
-
Potamianos, G.1
Neti, C.2
Gravier, G.3
Garg, A.4
Senior, A.5
-
28
-
-
26944441598
-
Data-driven refinement of a probabilistic model of user affect
-
C. Conati and H. Mclaren, L. Ardissono, P. Brna, and A. Mitrovic, Eds, Berlin, Germany
-
C. Conati and H. Mclaren, L. Ardissono, P. Brna, and A. Mitrovic, Eds., "Data-driven refinement of a probabilistic model of user affect," in Proc. 10th Int. Conf. User Modeling (UM2005), Berlin, Germany, 2005, pp. 40-49.
-
(2005)
Proc. 10th Int. Conf. User Modeling (UM2005)
, pp. 40-49
-
-
-
29
-
-
2442456966
-
A multilayer personality model
-
Hawthorne, NY, Jun
-
S. Kshirsagar and N. Magnenat-Thalmann, "A multilayer personality model," in Proc. 2nd Int. Symp. Smart Graphics (SMARTGRAPH 02), Hawthorne, NY, Jun. 2002, pp. 107-115.
-
(2002)
Proc. 2nd Int. Symp. Smart Graphics (SMARTGRAPH 02)
, pp. 107-115
-
-
Kshirsagar, S.1
Magnenat-Thalmann, N.2
-
30
-
-
33646752807
-
Learning dynamic audio-visual mapping with input-output midden Markov models
-
Jun
-
Y. Li and H. Shum, "Learning dynamic audio-visual mapping with input-output midden Markov models," IEEE Trans. Multimedia, vol. 8, no. 3, pp. 542-549, Jun. 2006.
-
(2006)
IEEE Trans. Multimedia
, vol.8
, Issue.3
, pp. 542-549
-
-
Li, Y.1
Shum, H.2
-
31
-
-
48149096367
-
Measurements of articulatory variations and communicative signals in expressive speech
-
Sep
-
M. Nordstrand, G. Svanfeldt, B. Granström, and D. House, "Measurements of articulatory variations and communicative signals in expressive speech," in Audio Visual Speech Process. (AVSP 03), Sep. 2003, pp. 233-237.
-
(2003)
Audio Visual Speech Process. (AVSP 03)
, pp. 233-237
-
-
Nordstrand, M.1
Svanfeldt, G.2
Granström, B.3
House, D.4
-
32
-
-
64149104077
-
Coproduction of speech and emotions: Visual and acoustic modifications of some phonetic labial targets
-
Sep
-
E. M. Caldognetto, P. Cosi, C. Drioli, G. Tisato, and F. Cavicchio, "Coproduction of speech and emotions: Visual and acoustic modifications of some phonetic labial targets," in Audio Visual Speech Process. (AVSP 03), Sep. 2003, pp. 209-214.
-
(2003)
Audio Visual Speech Process. (AVSP 03)
, pp. 209-214
-
-
Caldognetto, E.M.1
Cosi, P.2
Drioli, C.3
Tisato, G.4
Cavicchio, F.5
-
34
-
-
0003959340
-
Affective Computing MIT Media Laboratory Perceptual Computing Section, Cambridge, MA
-
Tech. Rep. 321
-
R. W. Picard, Affective Computing MIT Media Laboratory Perceptual Computing Section, Cambridge, MA, 1995, Tech. Rep. 321.
-
(1995)
-
-
Picard, R.W.1
-
35
-
-
64149120647
-
Praat, a System for Doing Phonetics by Computer Inst. Phonetic Sci., Univ. Amsterdam, Amsterdam, The Netherlands
-
Tech. Rep. 132 [Online, Available
-
P. Boersma and D. Weeninck, Praat, a System for Doing Phonetics by Computer Inst. Phonetic Sci., Univ. Amsterdam, Amsterdam, The Netherlands, 1996, Tech. Rep. 132 [Online]. Available: http://www. praat.org
-
(1996)
-
-
Boersma, P.1
Weeninck, D.2
-
36
-
-
64149118512
-
Predicting face movements from speech acoustics using spectral dynamics
-
Lausanne, Switzerland, Aug
-
J. Jiang, A. Alwan, L. Bernstein, E. Auer, Jr, and P. Keating, "Predicting face movements from speech acoustics using spectral dynamics," in Proc. IEEE Int. Conf. Multimedia and Expo (ICME 02), Lausanne, Switzerland, Aug. 2002, vol. 1, pp. 181-184.
-
(2002)
Proc. IEEE Int. Conf. Multimedia and Expo (ICME 02)
, vol.1
, pp. 181-184
-
-
Jiang, J.1
Alwan, A.2
Bernstein, L.3
Auer Jr, E.4
Keating, P.5
-
37
-
-
84890517975
-
Least-squares fitting of two 3-D pointsets
-
Sep
-
K. Arun, T. Huang, and S. Blostein, "Least-squares fitting of two 3-D pointsets," IEEE Trans. Pattern Anal. Mach. Intell., vol. 9, no. 5, pp. 698-700, Sep. 1987.
-
(1987)
IEEE Trans. Pattern Anal. Mach. Intell
, vol.9
, Issue.5
, pp. 698-700
-
-
Arun, K.1
Huang, T.2
Blostein, S.3
-
38
-
-
0036650155
-
Speaking mode variability in multimodal speech production
-
Jul
-
E. Vatikiotis-Bateson and H. C. Yehia, "Speaking mode variability in multimodal speech production," IEEE Trans. Neural Netw., vol. 13, no. 4, pp. 894-899, Jul. 2002.
-
(2002)
IEEE Trans. Neural Netw
, vol.13
, Issue.4
, pp. 894-899
-
-
Vatikiotis-Bateson, E.1
Yehia, H.C.2
-
39
-
-
64149131588
-
-
S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw, X. Liu, G. Moore, J. Odell, D. Ollason, D. Povey, V. Valtchev, and P. Woodland, The HTK Book. Cambridge, U.K.: Entropic Cambridge Res. Lab., 2006.
-
S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw, X. Liu, G. Moore, J. Odell, D. Ollason, D. Povey, V. Valtchev, and P. Woodland, The HTK Book. Cambridge, U.K.: Entropic Cambridge Res. Lab., 2006.
-
-
-
-
40
-
-
34047263009
-
Visual model structures and synchrony constraints for audio-visual speech recognition
-
May
-
T. Hazen, "Visual model structures and synchrony constraints for audio-visual speech recognition," IEEE Trans. Audio, Speech, Lang. Process., vol. 14, no. 3, pp. 1082-1089, May 2006.
-
(2006)
IEEE Trans. Audio, Speech, Lang. Process
, vol.14
, Issue.3
, pp. 1082-1089
-
-
Hazen, T.1
-
41
-
-
48149084430
-
Interplay between linguistic and affective goalsin facial expression during emotional utterances
-
Ubatuba-SP, Brazil, Dec
-
C. Busso and S. Narayanan, "Interplay between linguistic and affective goalsin facial expression during emotional utterances," in Proc. 7th Int. Seminar Speech Prod. (ISSP 06), Ubatuba-SP, Brazil, Dec. 2006, pp. 549-556.
-
(2006)
Proc. 7th Int. Seminar Speech Prod. (ISSP 06)
, pp. 549-556
-
-
Busso, C.1
Narayanan, S.2
-
42
-
-
15744397790
-
Mind reading machines: Automated inference ofcognitive mental states from video
-
The Hague, The Netherlands, Oct
-
R. E. Kaliouby and P. Robinson, "Mind reading machines: Automated inference ofcognitive mental states from video," in IEEE Conf. Syst., Man, Cybern., The Hague, The Netherlands, Oct. 2004, vol. 1, pp. 682-688.
-
(2004)
IEEE Conf. Syst., Man, Cybern
, vol.1
, pp. 682-688
-
-
Kaliouby, R.E.1
Robinson, P.2
-
43
-
-
15744380201
-
Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior
-
The Hague, The Netherlands, Oct
-
J. Cohn, L. Reed, Z. Ambadar, J. Xiao, and T. Moriyama, "Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior," in Proc. IEEE Conf. Syst., Man, Cybern., The Hague, The Netherlands, Oct. 2004, vol. 1, pp. 610-616.
-
(2004)
Proc. IEEE Conf. Syst., Man, Cybern
, vol.1
, pp. 610-616
-
-
Cohn, J.1
Reed, L.2
Ambadar, Z.3
Xiao, J.4
Moriyama, T.5
-
44
-
-
0033062198
-
Attention to facial regions in segmental and prosodic visual speech perception tasks
-
Jun
-
C. Lansing and G. McConkie, "Attention to facial regions in segmental and prosodic visual speech perception tasks," J. Speech, Lang., Hearing Res., vol. 42, pp. 526-539, Jun. 1999.
-
(1999)
J. Speech, Lang., Hearing Res
, vol.42
, pp. 526-539
-
-
Lansing, C.1
McConkie, G.2
-
45
-
-
44949119712
-
The importance of different facial areas for signalling visual prominence
-
Pittsburgh, PA, Sep
-
M. Swerts and E. Krahmer, "The importance of different facial areas for signalling visual prominence," in Proc. Int. Conf. Spoken Lang. Process. (ICSLP), Pittsburgh, PA, Sep. 2006, pp. 1280-1283.
-
(2006)
Proc. Int. Conf. Spoken Lang. Process. (ICSLP)
, pp. 1280-1283
-
-
Swerts, M.1
Krahmer, E.2
|