메뉴 건너뛰기




Volumn 124, Issue 5, 2008, Pages 3183-3190

A linear model of acoustic-to-facial mapping: Model parameters, data set size, and generalization across speakers

Author keywords

[No Author keywords available]

Indexed keywords

ACOUSTIC SIGNALS; ACOUSTIC WINDOWS; AUDIO VISUALS; CRITICAL SIZES; DATA SETS; FACIAL MAPPINGS; FACIAL MOTIONS; LINEAR MODELS; LINEAR TRANSFORMATIONS; MODEL PARAMETERS; RECORDED MOTIONS; SPEECH PERCEPTIONS; TRAINING SETS; VISUAL ASPECTS; VISUAL SPEECHES; WINDOW SIZES;

EID: 56749174163     PISSN: 00014966     EISSN: None     Source Type: Journal    
DOI: 10.1121/1.2982369     Document Type: Article
Times cited : (10)

References (34)
  • 2
    • 0012066381 scopus 로고    scopus 로고
    • Spatial and temporal variability in gestural specification
    • in, edited by W. Hulstijn, F. Peters, and P. van Lieshout (Elsevier Science, Amsterdam)
    • Alfonso, P. J., and Van Lieshout, P. (1997). " Spatial and temporal variability in gestural specification.," in Speech Production: Motor Control, Brain Research and Fluency Disorders, edited by, W. Hulstijn, F. Peters, and, P. van Lieshout, (Elsevier Science, Amsterdam), pp. 151-160.
    • (1997) Speech Production: Motor Control, Brain Research and Fluency Disorders , pp. 151-160
    • Alfonso, P.J.1    Van Lieshout, P.2
  • 3
    • 0017968519 scopus 로고
    • Inversion of articulatory-to-acoustic transformation in the vocal tract by a computer sorting technique
    • ".
    • Atal, B. S., Chang, J. J., Mathews, M. V., and Tukey, J. W. (1978). " Inversion of articulatory-to-acoustic transformation in the vocal tract by a computer sorting technique.," J. Acoust. Soc. Am. 63, 1535-1555.
    • (1978) J. Acoust. Soc. Am. , vol.63 , pp. 1535-1555
    • Atal, B.S.1    Chang, J.J.2    Mathews, M.V.3    Tukey, J.W.4
  • 8
    • 34447624135 scopus 로고    scopus 로고
    • Suitability of a UV-based video recording system for the analysis of small facial motions during speech
    • ".
    • Craig, M., Van Lieshout, P., and Wong, W. (2007). " Suitability of a UV-based video recording system for the analysis of small facial motions during speech.," Speech Commun. 49, 679-686.
    • (2007) Speech Commun. , vol.49 , pp. 679-686
    • Craig, M.1    Van Lieshout, P.2    Wong, W.3
  • 12
    • 56749105261 scopus 로고    scopus 로고
    • Lip and jaw coarticulation
    • in, edited by W. Hardcastle and N. Hewlett (Cambridge University Press, Cambridge)
    • Fletcher, J., and Harrington, J. (1999). " Lip and jaw coarticulation.," in Coarticulation: Theory, Data and Techniques, edited by, W. Hardcastle, and, N. Hewlett, (Cambridge University Press, Cambridge), pp. 164-178.
    • (1999) Coarticulation: Theory, Data and Techniques , pp. 164-178
    • Fletcher, J.1    Harrington, J.2
  • 14
    • 0034113506 scopus 로고    scopus 로고
    • Lip-jaw and tongue-jaw coordination during rate-controlled syllable repetitions
    • Hertrick, I., and Ackermann, H. (2000). " Lip-jaw and tongue-jaw coordination during rate-controlled syllable repetitions.," J. Acoust. Soc. Am. 107, 2236-2246.
    • (2000) J. Acoust. Soc. Am. , vol.107 , pp. 2236-2246
    • Hertrick, I.1    Ackermann, H.2
  • 15
    • 34247647975 scopus 로고    scopus 로고
    • Inverting mappings from smooth paths through Rn to paths through Rm: A technique applied to recovering articulation from acoustics
    • ".
    • Hogden, J., Rubin, P., McDermott, E., Katagiri, S., and Goldstein, L. (2007). " Inverting mappings from smooth paths through Rn to paths through Rm: A technique applied to recovering articulation from acoustics.," Speech Commun. 49, 361-383.
    • (2007) Speech Commun. , vol.49 , pp. 361-383
    • Hogden, J.1    Rubin, P.2    McDermott, E.3    Katagiri, S.4    Goldstein, L.5
  • 16
    • 0036650837 scopus 로고    scopus 로고
    • Real-time speech-driven face animation with expressions using neural networks
    • ".
    • Hong, P., Wen, Z., and Huang, T. S. (2002). " Real-time speech-driven face animation with expressions using neural networks.," IEEE Trans. Neural Netw. 13, 916-927.
    • (2002) IEEE Trans. Neural Netw. , vol.13 , pp. 916-927
    • Hong, P.1    Wen, Z.2    Huang, T.S.3
  • 18
    • 56749180289 scopus 로고    scopus 로고
    • " Proceedings of the 5th International Conference on Signal Processing, Beijing
    • Jiang, T., Li, Y., and Chen, H. (2000b). " A 1.44 Kbps vocoder based on LSP.," Proceedings of the 5th International Conference on Signal Processing, Beijing, pp. 697-701.
    • (2000) A 1.44 Kbps Vocoder Based on LSP , pp. 697-701
    • Jiang, T.1    Li, Y.2    Chen, H.3
  • 20
    • 50249158527 scopus 로고    scopus 로고
    • Version 2. Technical report, Department Electrical and Computer Engineering, McGill University, Montreal.
    • Kabal, P. (2003). " Time windows for linear prediction of speech.," Version 2. Technical report, Department Electrical and Computer Engineering, McGill University, Montreal.
    • (2003) Time Windows for Linear Prediction of Speech
    • Kabal, P.1
  • 22
    • 33747766904 scopus 로고    scopus 로고
    • A comparison of acoustic coding models for speech-driven facial animation
    • ".
    • Kakumanu, P., Esposito, A., Garcia, O., and Gutierrez-Osuna, R. (2006). " A comparison of acoustic coding models for speech-driven facial animation.," Speech Commun. 48, 598-615.
    • (2006) Speech Commun. , vol.48 , pp. 598-615
    • Kakumanu, P.1    Esposito, A.2    Garcia, O.3    Gutierrez-Osuna, R.4
  • 24
    • 0024344665 scopus 로고
    • Segmental intelligibility of synthetic speech produced by rule
    • ".
    • Logan, J. S., Greene, B. G., and Pisoni, D. B. (1989). " Segmental intelligibility of synthetic speech produced by rule.," J. Acoust. Soc. Am. 86, 566-581.
    • (1989) J. Acoust. Soc. Am. , vol.86 , pp. 566-581
    • Logan, J.S.1    Greene, B.G.2    Pisoni, D.B.3
  • 26
    • 33745712098 scopus 로고    scopus 로고
    • Speaker-independent 3D face synthesis driven by speech and text
    • ".
    • Savran, A., Arslan, L., and Akarun, L. (2006). " Speaker-independent 3D face synthesis driven by speech and text.," Signal Process. 86, 2932-2951.
    • (2006) Signal Process. , vol.86 , pp. 2932-2951
    • Savran, A.1    Arslan, L.2    Akarun, L.3
  • 27
    • 0014077928 scopus 로고
    • Determination of the geometry of the human vocal tract by acoustic measurements
    • Schroeder, M. R. (1967). " Determination of the geometry of the human vocal tract by acoustic measurements.," J. Acoust. Soc. Am. 41, 1002-1010.
    • (1967) J. Acoust. Soc. Am. , vol.41 , pp. 1002-1010
    • Schroeder, M.R.1
  • 28
    • 0000352807 scopus 로고
    • Trade-oils in tongue, jaw and palate contributions to speech production
    • Stone, M., and Vatikiotis-Bateson, E. (1995). " Trade-oils in tongue, jaw and palate contributions to speech production.," J. Phonetics 23, 81-100.
    • (1995) J. Phonetics , vol.23 , pp. 81-100
    • Stone, M.1    Vatikiotis-Bateson, E.2
  • 29
    • 0027128576 scopus 로고
    • Lipreading and audio-visual speech perception
    • Summerfield, Q. (1992). " Lipreading and audio-visual speech perception.," Philos. Trans. R. Soc. London, Ser. B 335 (1273), 71-78.
    • (1992) Philos. Trans. R. Soc. London, Ser. B , vol.335 , Issue.1273 , pp. 71-78
    • Summerfield, Q.1
  • 30
    • 34147186624 scopus 로고    scopus 로고
    • A coupled HMM approach to video-realistic speech animation
    • Xie, L., and Liu, Z.-Q. (2007). " A coupled HMM approach to video-realistic speech animation.," Pattern Recogn. Lett. 40, 2325-2340.
    • (2007) Pattern Recogn. Lett. , vol.40 , pp. 2325-2340
    • Xie, L.1    Liu, Z.-Q.2
  • 31
    • 0032179320 scopus 로고    scopus 로고
    • Lip movement synthesis from speech based on hidden Markov models
    • ".
    • Yamamoto, E., Nakamura, S., and Shikano, K. (1998). " Lip movement synthesis from speech based on hidden Markov models.," Speech Commun. 26, 105-115.
    • (1998) Speech Commun. , vol.26 , pp. 105-115
    • Yamamoto, E.1    Nakamura, S.2    Shikano, K.3
  • 32
    • 0032178592 scopus 로고    scopus 로고
    • Quantitative association of vocal-tract and facial behavior
    • ".
    • Yehia, H., Rubin, P., and Vatikiotis-Bateson, E. (1998). " Quantitative association of vocal-tract and facial behavior.," Speech Commun. 26, 23-43.
    • (1998) Speech Commun. , vol.26 , pp. 23-43
    • Yehia, H.1    Rubin, P.2    Vatikiotis-Bateson, E.3
  • 33
    • 0036656895 scopus 로고    scopus 로고
    • Linking facial animation, head motion and speech acoustics
    • ".
    • Yehia, H., Kuratate, T., and Vatikiotis-Bateson, E. (2001). " Linking facial animation, head motion and speech acoustics.," J. Phonetics 30, 555-568.
    • (2001) J. Phonetics , vol.30 , pp. 555-568
    • Yehia, H.1    Kuratate, T.2    Vatikiotis-Bateson, E.3
  • 34
    • 33749437734 scopus 로고    scopus 로고
    • Design, implementation and evaluation of the Czech realistic audio-visual speech synthesis
    • ".
    • Zelezny, M., Krnoul, Z., Cisar, P., and Matousek, J. (2006). " Design, implementation and evaluation of the Czech realistic audio-visual speech synthesis.," Signal Process. 86, 3657-3673.
    • (2006) Signal Process. , vol.86 , pp. 3657-3673
    • Zelezny, M.1    Krnoul, Z.2    Cisar, P.3    Matousek, J.4


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.