메뉴 건너뛰기




Volumn , Issue , 2008, Pages 2318-2321

A trainable trajectory formation model TD-HMM parameterized for the LIPS 2008 challenge

Author keywords

Facial animation audiovisual speech synthesis; HMM

Indexed keywords

ASYNCHRONY; AUDIO-VISUAL SPEECH; CONTEXT DEPENDENT; HMM; MOTION CAPTURE DATA; PARAMETERIZED; PREDICTION ERRORS; TRAJECTORY FORMATION; TRIPHONES;

EID: 84867216889     PISSN: None     EISSN: 19909772     Source Type: Conference Proceeding    
DOI: None     Document Type: Conference Paper
Times cited : (3)

References (23)
  • 1
    • 0036656541 scopus 로고    scopus 로고
    • Three-dimensional linear articulatory modeling of tongue, lips and face based on MRI and video images
    • Badin, P., et al., Three-dimensional linear articulatory modeling of tongue, lips and face based on MRI and video images. Journal of Phonetics, 2002. 30(3): p. 533-553.
    • (2002) Journal of Phonetics , vol.30 , Issue.3 , pp. 533-553
    • Badin, P.1
  • 5
    • 84883424118 scopus 로고
    • Rule-based Visual Speech Synthesis
    • Madrid, Spain
    • Beskow, J. Rule-based Visual Speech Synthesis. in Eurospeech. 1995. Madrid, Spain. p. 299-302.
    • (1995) Eurospeech , pp. 299-302
    • Beskow, J.1
  • 6
    • 0001514782 scopus 로고
    • Modeling coarticulation in synthetic visual speech
    • D. Thalmann and N. Magnenat-Thalmann, Editors. Springer-Verlag: Tokyo
    • Cohen, M.M. and D.W. Massaro, Modeling coarticulation in synthetic visual speech, in Models and Techniques in Computer Animation, D. Thalmann and N. Magnenat-Thalmann, Editors. 1993, Springer-Verlag: Tokyo. p. 141-155.
    • (1993) Models and Techniques in Computer Animation , pp. 141-155
    • Cohen, M.M.1    Massaro, D.W.2
  • 7
    • 0019256553 scopus 로고
    • The detection of audiovisual desynchrony
    • Dixon, N.F. and L. Spitz, The detection of audiovisual desynchrony. Perception, 1980. 9: p. 719-721.
    • (1980) Perception , vol.9 , pp. 719-721
    • Dixon, N.F.1    Spitz, L.2
  • 8
    • 0004131347 scopus 로고    scopus 로고
    • Univ. Eng. Dept. University of Cambridge: Cambridge, UK
    • Donovan, R., Trainable speech synthesis, in Univ. Eng. Dept. 1996, University of Cambridge: Cambridge, UK p. 164.
    • (1996) Trainable Speech Synthesis , pp. 164
    • Donovan, R.1
  • 11
    • 44949159884 scopus 로고    scopus 로고
    • TDA: A new trainable trajectory formation system for facial animation
    • Pittsburgh, PE
    • Govokhina, O., et al. TDA: A new trainable trajectory formation system for facial animation. in InterSpeech. 2006. Pittsburgh, PE. p. 2474-2477.
    • (2006) InterSpeech , pp. 2474-2477
    • Govokhina, O.1
  • 13
    • 78649309580 scopus 로고    scopus 로고
    • Visual model structures and synchrony constraints for audio-visual speech recognition
    • Hazen, T.J., Visual model structures and synchrony constraints for audio-visual speech recognition. IEEE Trans. on Speech and Audio Processing, 2005.
    • (2005) IEEE Trans. on Speech and Audio Processing
    • Hazen, T.J.1
  • 14
    • 0030004909 scopus 로고    scopus 로고
    • A model of articulator trajectory formation based on the motor tasks of vocal-tract shapes
    • Kaburagi, T. and M. Honda, A model of articulator trajectory formation based on the motor tasks of vocal-tract shapes. Journal of the Acoustical Society of America, 1996. 99(5): p. 3154-3170.
    • (1996) Journal of the Acoustical Society of America , vol.99 , Issue.5 , pp. 3154-3170
    • Kaburagi, T.1    Honda, M.2
  • 15
    • 85034718268 scopus 로고    scopus 로고
    • Audio-visual synthesis of talking faces from speech production correlates
    • Kuratate, T., et al. Audio-visual synthesis of talking faces from speech production correlates. in EuroSpeech. 1999. p. 1279-1282.
    • (1999) EuroSpeech , pp. 1279-1282
    • Kuratate, T.1
  • 16
    • 0017199877 scopus 로고
    • Hearing lips and seeing voices
    • McGurk, H. and J. MacDonald, Hearing lips and seeing voices. Nature, 1976. 264: p. 746-748.
    • (1976) Nature , vol.264 , pp. 746-748
    • McGurk, H.1    MacDonald, J.2
  • 18
    • 84919370414 scopus 로고    scopus 로고
    • Text-to-audio-visual speech synthesis based on parameter generation from HMM
    • Budapest, Hungary
    • Tamura, M., et al. Text-to-audio-visual speech synthesis based on parameter generation from HMM. in EUROSPEECH. 1999. Budapest, Hungary. p. 959-962.
    • (1999) EUROSPEECH , pp. 959-962
    • Tamura, M.1
  • 19
    • 85133460248 scopus 로고    scopus 로고
    • Visual speech synthesis based on parameter generation from HMM: Speech-driven and text-and-speech-driven approaches
    • Tamura, M., et al. Visual speech synthesis based on parameter generation from HMM: speech-driven and text-and-speech-driven approaches. in Auditory-visual Speech Processing Workshop. 1998. Terrigal, Sydney, Australia. p. 219-224.
    • Auditory-visual Speech Processing Workshop. 1998. Terrigal, Sydney, Australia , pp. 219-224
    • Tamura, M.1
  • 22
    • 84936526808 scopus 로고
    • Coarticulation is largely planned
    • Whalen, D.H., Coarticulation is largely planned. Journal of Phonetics, 1990. 18(1): p. 3-35.
    • (1990) Journal of Phonetics , vol.18 , Issue.1 , pp. 3-35
    • Whalen, D.H.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.