{"title":"Effect of Visual Speech in Sign Speech Synthesis","authors":"Zdenek Krnoul","volume":31,"journal":"International Journal of Computer and Information Engineering","pagesStart":1671,"pagesEnd":1677,"ISSN":"1307-6892","URL":"https:\/\/publications.waset.org\/pdf\/4362","abstract":"
This article investigates a contribution of synthesized visual speech. Synthesis of visual speech expressed by a computer consists in an animation in particular movements of lips. Visual speech is also necessary part of the non-manual component of a sign language. Appropriate methodology is proposed to determine the quality and the accuracy of synthesized visual speech. Proposed methodology is inspected on Czech speech. Hence, this article presents a procedure of recording of speech data in order to set a synthesis system as well as to evaluate synthesized speech. Furthermore, one option of the evaluation process is elaborated in the form of a perceptual test. This test procedure is verified on the measured data with two settings of the synthesis system. The results of the perceptual test are presented as a statistically significant increase of intelligibility evoked by real and synthesized visual speech. Now, the aim is to show one part of evaluation process which leads to more comprehensive evaluation of the sign speech synthesis system.<\/p>\r\n","references":"[1] R. Conrad, The deaf school child. London: Harper & Row, 1979.\r\n[2] O. Velehradsk'a and K. Kuchler, \"Pr \u252c\u2591uzkum \u2566\u00e7cten'a\u2566\u00e7rsk'ych dovednost'\u2500\u2592 na\r\n\u2566\u00e7skol'ach pro d\u2566\u00e7eti s vadami sluchu,\" INFO-Zpravodaj FRPSP, vol. 6,\r\n1998.\r\n[3] P. Campr, M. Hr 'uz, A. Karpov, P. Santemiz, M. \u2566\u00e7Zelezn'y,\r\nand O. Aran, \"Sign-language-enabled information kiosk,\" 2009.\r\n(Online). Available: http:\/\/www.kky.zcu.cz\/en\/publications\/CamprP\r\n2009 Sign-language-enabled\r\n[4] M. \u2566\u00e7Zelezn'y, Z. Kr \u2566\u00e7noul, P. C'\u2500\u2592sa\u2566\u00e7r, and J. Matou\u2566\u00e7sek, \"Design, implementation\r\nand evaluation of the czech realistic audio-visual speech synthesis,\"\r\nSignal Procesing, Special section: Multimodal human-computer interfaces,\r\nvol. 86, pp. 3657-3673, 2006.\r\n[5] V. Radov'a and P. Vop'alka, \"Methods of sentences selection for readspeech\r\ncorpus design,\" Lecture Notes In Computer Science, vol. 1692,\r\n1999.\r\n[6] J. Psutka, L. M\u252c\u00bfuller, J. Matou\u2566\u00e7sek, and V. Radov'a, Mluv'\u2500\u2592me s po\u2566\u00e7c'\u2500\u2592ta\u2566\u00e7cem\r\n\u2566\u00e7cesky, 1st ed. Praha: Academia, 2006.\r\n[7] A. MacLeod and Q. Summerfield, \"A procedure for measuring auditory\r\nand audio-visual speech-reception thresholds for sentences in noise:\r\nrationale, evaluation, and recommendations for use,\" British Journal of\r\nAudiology, 24(1), 29-43, 1990.\r\n[8] A. B\u252c\u00bfohmov'a, J. Haji\u2566\u00e7c, E. Haji\u2566\u00e7cov'a, and B. Hladk'a, \"The prague dependency\r\ntreebank: Three-level annotation scenario,\" Treebanks: Building\r\nand Using Syntactically Annotated Corpora, ed. Anne Abeille. Kluwer\r\nAcademic Publishers, 2001.\r\n[9] M. M. Cohen and D. W. Massaro, \"Modeling coarticulation in synthetic\r\nvisual speech,\" in Models and Techniques in Computer Animation, N.\r\nM. Thalmann & D. Thalmann, Ed. Tokyo: Springer-Verlag, 1993.\r\n[10] Z. Kr \u2566\u00e7noul and M. \u2566\u00e7Zelezn'y, \"Development of czech talking head,\" in\r\nProceedings of Interspeech 2008, Brisbane, Australia, 2008.\r\n[11] J. Beskow, \"Trainable articulatory control models for visual speech\r\nsynthesis,\" International Journal of Speech Technology, 2004, submitted.\r\n[12] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classification\r\nand Regression Trees, 1st ed. Chapman and Hall, Boca Raton, 1998.\r\n[13] Z. Kr \u2566\u00e7noul and M. \u2566\u00e7Zelezn'y, \"Realistic face animation for a Czech Talking\r\nHead,\" in Proceedings of TEXT, SPEECH and DIALOGUE, TSD 2004,\r\nBrno, Czech republic, 2004.\r\n[14] Z. Kr \u2566\u00e7noul, M. \u2566\u00e7Zelezn'y, P. C'\u2500\u2592sa\u2566\u00e7r, and J. Holas, \"Viseme analysis for\r\nspeech-driven facial animation for czech audio-visual speech synthesis,\"\r\nin Proceedings of SPECOM 2005, University of Patras, Greece, 2005.\r\n[15] Z. Kr \u2566\u00e7noul, P. C'\u2500\u2592sa\u2566\u00e7r, and M. \u2566\u00e7Zelezn'y, \"Face model reconstruction for\r\nczech audio-visual speech synthesis,\" in SPECOM 2004, St. Petersburg,\r\nRussian Federation, 2004.","publisher":"World Academy of Science, Engineering and Technology","index":"Open Science Index 31, 2009"}