메뉴 건너뛰기




Volumn , Issue , 2013, Pages 7952-7956

Articulatory inversion and synthesis: Towards articulatory-based modification of speech

Author keywords

articulatory inversion; articulatory synthesis; Maeda parameters; speech modification

Indexed keywords

ACOUSTIC DOMAINS; ARTICULATORY INVERSION; ARTICULATORY SYNTHESIS; CEPSTRAL; INVERSION MODELS; MAEDA PARAMETERS;

EID: 84890528210     PISSN: 15206149     EISSN: None     Source Type: Conference Proceeding    
DOI: 10.1109/ICASSP.2013.6639213     Document Type: Conference Paper
Times cited : (14)

References (22)
  • 1
    • 0024879199 scopus 로고
    • The effective second formant F2'and the vocal tract front-cavity
    • H. Hermansky and D. Broad, "The effective second formant F2'and the vocal tract front-cavity, " in ICASSP, 1989, pp. 480-483.
    • (1989) ICASSP , pp. 480-483
    • Hermansky, H.1    Broad, D.2
  • 2
    • 85027461504 scopus 로고    scopus 로고
    • Using articulatory position data in voice transformation
    • A. Toth and A. Black, "Using articulatory position data in voice transformation, " ISCA SSW6, pp. 182-187, 2007.
    • (2007) ISCA SSW6 , pp. 182-187
    • Toth, A.1    Black, A.2
  • 6
    • 38649140222 scopus 로고    scopus 로고
    • Statistical mapping between articulatory movements and acoustic spectrum using a Gaussian mixture model
    • T. Toda, A. W. Black, and K. Tokuda, "Statistical mapping between articulatory movements and acoustic spectrum using a Gaussian mixture model, " Speech Communication, vol. 50, pp. 215-227, 2008.
    • (2008) Speech Communication , vol.50 , pp. 215-227
    • Toda, T.1    Black, A.W.2    Tokuda, K.3
  • 7
    • 85008009610 scopus 로고    scopus 로고
    • Estimation of articulatory trajectories based on gaussian mixture model (gmm) with audio-visual information fusion and dynamic kalman smoothing
    • I. Y. Ozbek, M. Hasegawa-Johnson, and M. Demirekler, "Estimation of Articulatory Trajectories Based on Gaussian Mixture Model (GMM) with Audio-Visual Information Fusion and Dynamic Kalman Smoothing, " IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, pp. 1180-1195, 2011.
    • (2011) Ieee Transactions on Audio, Speech, and Language Processing , vol.19 , pp. 1180-1195
    • Ozbek, I.Y.1    Hasegawa-Johnson, M.2    Demirekler, M.3
  • 8
    • 70450219528 scopus 로고    scopus 로고
    • Preliminary inversion mapping results with a new EMA corpus
    • K. Richmond, "Preliminary inversion mapping results with a new EMA corpus, " in Proc. Interspeech, 2009, pp. 2835-2838.
    • (2009) Proc. Interspeech , pp. 2835-2838
    • Richmond, K.1
  • 9
    • 0036642567 scopus 로고    scopus 로고
    • Combining acoustic and articulatory feature information for robust speech recognition
    • K. Kirchhoff, G. A. Fink, and G. Sagerer, "Combining acoustic and articulatory feature information for robust speech recognition, " Speech Communication, vol. 37, pp. 303-319, 2002.
    • (2002) Speech Communication , vol.37 , pp. 303-319
    • Kirchhoff, K.1    Fink, G.A.2    Sagerer, G.3
  • 10
    • 79959813685 scopus 로고    scopus 로고
    • Robust word recognition using articulatory trajectories and Gestures
    • V. Mitra, H. Nam, C. Espy-Wilson, E. Saltzman, and L. Goldstein, "Robust word recognition using articulatory trajectories and Gestures, " in Interspeech, 2010, pp. 2038-2041.
    • (2010) Interspeech , pp. 2038-2041
    • Mitra, V.1    Nam, H.2    Espy-Wilson, C.3    Saltzman, E.4    Goldstein, L.5
  • 11
    • 82255163922 scopus 로고    scopus 로고
    • Automatic speech recognition using articulatory features from subject-independent acousticto-articulatory inversion
    • P. Ghosh and S. Narayanan, "Automatic speech recognition using articulatory features from subject-independent acousticto-articulatory inversion, " The Journal of the Acoustical Society of America, vol. 130, pp. EL251-EL257, 2011.
    • (2011) The Journal of the Acoustical Society of America , vol.130
    • Ghosh, P.1    Narayanan, S.2
  • 12
    • 34247634965 scopus 로고
    • An articulatory model of the tongue based on a statistical analysis
    • S. Maeda, "An articulatory model of the tongue based on a statistical analysis, " The Journal of the Acoustical Society of America, vol. 65, p. S22, 1979.
    • (1979) The Journal of the Acoustical Society of America , vol.65
    • Maeda, S.1
  • 13
    • 0020281396 scopus 로고
    • A digital simulation method of the vocal-tract system
    • S. Maeda, "A digital simulation method of the vocal-tract system, " Speech Communication, vol. 1, pp. 199-229, 1982.
    • (1982) Speech Communication , vol.1 , pp. 199-229
    • Maeda, S.1
  • 14
    • 34548248918 scopus 로고    scopus 로고
    • Training a vocal tract synthesizer to imitate speech using distal supervised learning
    • I. S. Howard and M. A. Huckvale, "Training a vocal tract synthesizer to imitate speech using distal supervised learning, " in Proc. SPECOM, 2005, pp. 159-162.
    • (2005) Proc. SPECOM , pp. 159-162
    • Howard, I.S.1    Huckvale, M.A.2
  • 15
    • 0031198820 scopus 로고    scopus 로고
    • Learning to speak. Sensori-motor control of speech movements
    • G. Bailly, "Learning to speak. Sensori-motor control of speech movements, " Speech Communication, vol. 22, pp. 251-267, 1997.
    • (1997) Speech Communication , vol.22 , pp. 251-267
    • Bailly, G.1
  • 18
    • 0030677481 scopus 로고    scopus 로고
    • Speech representation and transformation using adaptive interpolation of weighted spectrum: Vocoder revisited
    • H. Kawahara, "Speech representation and transformation using adaptive interpolation of weighted spectrum: vocoder revisited, " in ICASSP, 1997, pp. 1303-1306.
    • (1997) ICASSP , pp. 1303-1306
    • Kawahara, H.1
  • 19
    • 0029252331 scopus 로고
    • Nonlinear adaptive prediction of nonstationary signals
    • S. Haykin and L. Li, "Nonlinear adaptive prediction of nonstationary signals, " IEEE Transactions on Signal Processing, vol. 43, pp. 526-535, 1995.
    • (1995) IEEE Transactions on Signal Processing , vol.43 , pp. 526-535
    • Haykin, S.1    Li, L.2
  • 20
    • 57749193836 scopus 로고    scopus 로고
    • Voice conversion based on maximum-likelihood estimation of spectral parameter trajectory
    • T. Toda, A. W. Black, and K. Tokuda, "Voice conversion based on maximum-likelihood estimation of spectral parameter trajectory, " IEEE Transactions on Audio, Speech, and Language ProceSSing, vol. 15, pp. 2222-2235, 2007.
    • (2007) IEEE Transactions on Audio, Speech, and Language ProceSSing , vol.15 , pp. 2222-2235
    • Toda, T.1    Black, A.W.2    Tokuda, K.3
  • 21
    • 51449085174 scopus 로고    scopus 로고
    • Analysis-bysynthesis features for speech recognition
    • Z. AI Bawab, R. Bhiksha, and R. M. Stem, "Analysis-bysynthesis features for speech recognition, " in ICASSP, 2008, pp. 4185-4188.
    • (2008) ICASSP , pp. 4185-4188
    • Bawab, Z.A.I.1    Bhiksha, R.2    Stem, R.M.3
  • 22
    • 67650657780 scopus 로고    scopus 로고
    • Foreign accent conversion in computer assisted pronunciation training
    • D. Felps, H. Bortfeld, and R. Gutierrez-Osuna, "Foreign accent conversion in computer assisted pronunciation training, " Speech communication, vol. 51, pp. 920-932, 2009.
    • (2009) Speech Communication , vol.51 , pp. 920-932
    • Felps, D.1    Bortfeld, H.2    Gutierrez-Osuna, R.3


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.