Confusions Among Visually Perceived Consonants

1968; American Speech–Language–Hearing Association; Volume: 11; Issue: 4 Linguagem: Inglês

10.1044/jshr.1104.796

ISSN

0022-4685

Autores

Cletus G. Fisher,

Tópico(s)

Multisensory perception and integration

Resumo

No AccessJournal of Speech and Hearing ResearchResearch Article1 Dec 1968Confusions Among Visually Perceived Consonants Cletus G. Fisher Cletus G. Fisher University of Iowa, Iowa City, Iowa Google Scholar https://doi.org/10.1044/jshr.1104.796 SectionsAboutPDF ToolsAdd to favoritesDownload CitationTrack Citations ShareFacebookTwitterLinked In Eighteen college students with normal hearing responded to the visual perception of initial and final consonants in an English-like phonetic environment in a test of the homopheny of consonant sounds of English. The Multiple-choice Intelligibility Test provided stimulus items but special response sheets were provided to allow each subject a possible response of any consonant judged homotypical or homorganic to the stimulus item. Correct answers as possible responses were deleted to provide a usable number of confusions. Subjects were not aware of the deletion of correct responses even after the task was completed. Resulting confusion matrices were analyzed for significant confusions among consonants; these confusions were grouped into mutually exclusive classes termed visemes. The results tend to support previously published linguistic groupings of homophenous sounds rather than the classical listing from the developers of speechreading methodology. Variations from the former are explained in terms of the addition of minimal phonetic redundancy. Additional Resources FiguresReferencesRelatedDetailsCited by Brain and Language235 (105196)1 Dec 2022The timing of visual speech modulates auditory neural processingMarc Sato Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies6:3 (1-26)6 Sep 2022MuteItTanmay Srivastava, Prerna Khanna, Shijia Pan, Phuc Nguyen and Shubham Jain CAAI Transactions on Intelligence Technology17 Aug 2022Developing phoneme‐based lip‐reading sentences system for silent speech recognitionRanda El‐Bialy, Daqing Chen, Souheil Fenghour, Walid Hussein, Perry Xiao, Omar H. Karam and Bo Li Nergis Pervan Akman, Talya Tumer Sivri, Ali Berkol and Hamit Erdem (2022) Lip Reading Multiclass Classification by Using Dilated CNN with Turkish Dataset 2022 International Conference on Electrical, Computer and Energy Technologies (ICECET)10.1109/ICECET55527.2022.9873011978-1-6654-7087-2 Egyptian Informatics Journal1 Jul 2022Read my lips: Artificial intelligence word-level arabic lipreading systemWaleed Dweik, Sundus Altorman and Safa Ashour Cortex152 (21-35)1 Jul 2022Motor and visual influences on auditory neural processing during speaking and listeningMarc Sato American Journal of Audiology31:2 (453-469)2 Jun 2022Lipreading: A Review of Its Continuing Importance for Speech Recognition With an Acquired Hearing Loss and Possibilities for Effective TrainingLynne E. Bernstein, Nicole Jordan, Edward T. Auer and Silvio P. Eberhardt International Journal of Cognitive Computing in Engineering3 (24-30)1 Jun 2022Deep learning based assistive technology on audio visual speech recognition for hearing impairedL Ashok Kumar, D Karthika Renuka, S Lovelyn Rose, M C Shunmuga priya and I Made Wartana Proceedings of the ACM on Computer Graphics and Interactive Techniques5:1 (1-15)4 May 2022Joint Audio-Text Model for Expressive Speech-Driven 3D Facial AnimationYingruo Fan, Zhaojiang Lin, Jun Saito, Wenping Wang and Taku Komura NeuroImage252 (119044)1 May 2022Masking of the mouth area impairs reconstruction of acoustic speech features and higher-level segmentational features in the presence of a distractor speakerChandra Leon Haider, Nina Suess, Anne Hauswald, Hyojin Park and Nathan Weisz Psychonomic Bulletin & Review29:2 (600-612)1 Apr 2022The role of iconic gestures and mouth movements in face-to-face communicationAnna Krason, Rebecca Fenton, Rosemary Varley and Gabriella Vigliocco IEEE Internet of Things Journal9:7 (5357-5367)RealPRNet: A Real-Time Phoneme-Recognized Network for "Believable" Speech AnimationZixiao Yu, Haohong Wang and Jian Ren American Journal of Audiology31:1 (57-77)3 Mar 2022During Lipreading Training With Sentence Stimuli, Feedback Controls Learning and Generalization to Audiovisual Speech in NoiseLynne E. Bernstein, Edward T. Auer and Silvio P. Eberhardt Brain and Language225 (105058)1 Feb 2022How are visemes and graphemes integrated with speech sounds during spoken word recognition? ERP evidence for supra-additive responses during audiovisual compared to auditory speech processingChotiga Pattamadilok and Marc Sato IEEE/ACM Transactions on Audio, Speech, and Language Processing30 (2076-2090)End-to-End Lip-Reading Without Large-Scale DataAdriana Fernandez-Lopez and Federico M. Sukno Xian Liu, Yinghao Xu, Qianyi Wu, Hang Zhou, Wayne Wu and Bolei Zhou (2022) Semantic-Aware Implicit Neural Audio-Driven Video Portrait Generation Computer Vision – ECCV 202210.1007/978-3-031-19836-6_7 Sensors22:1 (72)23 Dec 2021Lipreading Architecture Based on Multiple Convolutional Neural Networks for Sentence-Level Visual Speech RecognitionSanghun Jeon, Ahmed Elsharkawy and Mun Sang Kim Souheil Fenghour, Daqing Chen, Laureta Hajderanj, Isakh Weheliye and Perry Xiao (2021) A Novel Supervised t-SNE Based Approach of Viseme Classification for Automated Lip Reading 2021 International Conference on Electrical, Computer and Energy Technologies (ICECET)10.1109/ICECET52533.2021.9698534978-1-6654-4231-2 Sensors21:23 (7890)26 Nov 2021An Effective Conversion of Visemes to Words for High-Performance Automatic LipreadingSouheil Fenghour, Daqing Chen, Kun Guo, Bo Li and Perry Xiao International Journal of Speech Technology25 Nov 2021Research on the application of speech recognition in computer network technology in the era of big dataBaohua Zhang Yudong Guo, Keyu Chen, Sen Liang, Yong-Jin Liu, Hujun Bao and Juyong Zhang (2021) AD-NeRF: Audio Driven Neural Radiance Fields for Talking Head Synthesis 2021 IEEE/CVF International Conference on Computer Vision (ICCV)10.1109/ICCV48922.2021.00573978-1-6654-2812-5 K.T. Bibish Kumar, Sunil John, K.M. Muraleedharan and R.K. Sunil Kumar (2021) Linguistically involved data-driven approach for Malayalam phoneme-to-viseme mapping Applied Speech Processing10.1016/B978-0-12-823898-1.00003-5 Zixiao Yu, Haohong Wang and Jian Ren (2020) A Hybrid Temporal Modeling Phoneme Recognized Network for Real-Time Speech Animation 2020 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR)10.1109/MIPR49039.2020.00019978-1-7281-4272-2 Cognitive Science44:71 Jul 2020Making Sense of the Hands and Mouth: The Role of "Secondary" Cues to Meaning in British Sign Language and EnglishPamela Perniss, David Vinson and Gabriella Vigliocco Navin Kumar Mudaliar, Kavita Hegde, Anand Ramesh and Varsha Patil (2020) Visual Speech Recognition: A Deep Learning Approach 2020 5th International Conference on Communication and Electronics Systems (ICCES)10.1109/ICCES48766.2020.9137926978-1-7281-5371-1 Bo Xu, Cheng Lu, Yandong Guo and Jacob Wang (2020) Discriminative Multi-Modality Speech Recognition 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR42600.2020.01444978-1-7281-7168-5 The Journal of the Acoustical Society of America147:4 (2609-2624)1 Apr 2020Multi-modal cross-linguistic perception of fricatives in clear speechSylvia Cho, Allard Jongman, Yue Wang and Joan A. Sereno Bo Xu, Jacob Wang, Cheng Lu and Yandong Guo (2020) Watch to Listen Clearly: Visual Speech Enhancement Driven Multi-modality Speech Recognition 2020 IEEE Winter Conference on Applications of Computer Vision (WACV)10.1109/WACV45572.2020.9093314978-1-7281-6553-0 IEEE Access8 (215516-215530)Lip Reading Sentences Using Deep Learning With Only Visual CuesSouheil Fenghour, Daqing Chen, Kun Guo and Perry Xiao Samuel Albanie, Gül Varol, Liliane Momeni, Triantafyllos Afouras, Joon Son Chung, Neil Fox and Andrew Zisserman (2020) BSL-1K: Scaling Up Co-articulated Sign Language Recognition Using Mouthing Cues Computer Vision – ECCV 202010.1007/978-3-030-58621-8_3 Frederic Chaume (2020) Dubbing The Palgrave Handbook of Audiovisual Translation and Media Accessibility10.1007/978-3-030-42105-2_6 Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies3:4 (1-24)11 Dec 2019RFID TattooJingxian Wang, Chengfeng Pan, Haojian Jin, Vaibhav Singh, Yash Jain, Jason I. Hong, Carmel Majidi and Swarun Kumar Behavior Research Methods10 Dec 2019Motion capture-based animated characters for the study of speech–gesture integrationJens Nirme, Magnus Haake, Agneta Gulz and Marianne Gullberg International Journal of Speech Technology22:4 (1149-1166)1 Dec 2019Viseme set identification from Malayalam phonemes and allophonesK. T. Bibish Kumar, R. K. Sunil Kumar, E. P. A. Sandesh, S. Sourabh and V. L. Lajish Developmental Science22:61 Nov 2019The role of production abilities in the perception of consonant category in infantsAnne Vilain, Marjorie Dole, Hélène Lœvenbruck, Olivier Pascalis and Jean‐Luc Schwartz Applied Sciences9:18 (3870)15 Sep 2019Alternative Visual Units for an Optimized Phoneme-Based Lipreading SystemHelen L. Bear and Richard Harvey Frederic Chaume (2019) Chapter 5. Audiovisual translation in the age of digital transformation Reassessing Dubbing10.1075/btl.148.05cha6 Aug 2019 Image and Vision Computing88 (76-83)1 Aug 2019Lip reading with Hahn Convolutional Neural NetworksAbderrahim Mesbah, Aissam Berrahou, Hicham Hammouchi, Hassan Berbia, Hassan Qjidaa and Mohamed Daoudi Language, Cognition and Neuroscience (1-18)16 Jul 2019The contribution of audiovisual speech to lexical-semantic processing in natural spoken sentencesAngèle Brunellière, Laurence Delrue and Cyril Auran Computer Speech & Language55 (101-119)1 May 2019Synthesising visual speech using dynamic visemes and deep learning architecturesAusdang Thangthai, Ben Milner and Sarah Taylor PLOS ONE14:3 (e0213588)21 Mar 2019Perception of incongruent audiovisual English consonantsKaylah Lalonde, Lynne A. Werner and Julie Jeannette Gros-Louis Frontiers in Communication37 Jan 2019Visual-Tactile Speech Perception and the Autism QuotientDonald Derrick, Katie Bicevskis and Bryan Gick Souheil Fenghour, Daqing Chen and Perry Xiao (2019) Decoder-Encoder LSTM for Lip Reading the 2019 8th International Conference10.1145/3328833.33288459781450361057 Adriana Fernandez-Lopez and Federico M. Sukno (2019) Optimizing Phoneme-to-Viseme Mapping for Continuous Lip-Reading in Spanish Computer Vision, Imaging and Computer Graphics – Theory and Applications10.1007/978-3-030-12209-6_15 International Journal of Audiology57:12 (914-922)2 Dec 2018Speech intelligibility of virtual humansAnnelies Devesse, Alexander Dudek, Astrid van Wieringen and Jan Wouters Computer Speech & Language52 (165-190)1 Nov 2018Comparing heterogeneous visual gestures for measuring the diversity of visual speech signalsHelen L. Bear and Richard Harvey Image and Vision Computing78 (53-72)1 Oct 2018Survey on automatic lip-reading in the era of deep learningAdriana Fernandez-Lopez and Federico M. Sukno S. Cygert, G. Szwoch, S. Zaporowski and A. Czyzewski (2018) Vocalic Segments Classification Assisted by Mouth Motion Capture 2018 11th International Conference on Human System Interaction (HSI)10.1109/HSI.2018.8430943978-1-5386-5024-0 Language Learning68 (127-158)1 Jun 2018Response Errors in Females' and Males' Sentence Lipreading Necessitate Structurally Different Models for Predicting Lipreading AccuracyLynne E. Bernstein E.P.A. Sandesh and V. L. Lajish (2018) Lip Motion Synthesis for Speech Animation Using Active Shape Model 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS)10.1109/ICCONS.2018.8663133978-1-5386-2842-3 Multisensory Research31:1-2 (57-78)The Time Course of Audio-Visual Phoneme Identification: a High Temporal Resolution StudyCarolina Sánchez-García, Sonia Kandel, Christophe Savariaux and Salvador Soto-Faraco Speech Communication95 (40-67)1 Dec 2017Phoneme-to-viseme mappings: the good, the bad, and the uglyHelen L Bear and Richard Harvey Communication Sciences & Disorders22:3 (615-628)30 Sep 2017Analysis of Korean Viseme System in Korean Standard Monosyllabic Word ListsJaehee Choi, Keonseok Yoon, Hyesoo Ryu and Hyunsook Jang Speech Communication92 (114-124)1 Sep 2017The influence of auditory-visual speech and clear speech on cross-language perceptual assimilationSarah E. Fenwick, Catherine T. Best, Chris Davis and Michael D. Tyler ACM Transactions on Graphics36:4 (1-12)20 Jul 2017Audio-driven facial animation by joint end-to-end learning of pose and emotionTero Karras, Timo Aila, Samuli Laine, Antti Herva and Jaakko Lehtinen Benjamin M. Gorman and David R. Flatla (2017) A Framework for Speechreading Acquisition Tools CHI '17: CHI Conference on Human Factors in Computing Systems10.1145/3025453.302556097814503465592 May 2017 Frontiers in Human Neuroscience1011 Jan 2017Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent LipreadingAisling E. O'Sullivan, Michael J. Crosse, Giovanni M. Di Liberto and Edmund C. Lalor The Journal of the Acoustical Society of America140:5 (3531-3539)1 Nov 2016Visual-tactile integration in speech perception: Evidence for modality neutral speech primitivesKatie Bicevskis, Donald Derrick and Bryan Gick Hyun-Jun Hyung, Byeong-Kyu Ahn, Dongwoon Choi, Dukyeon Lee and Dong-Wook Lee (2016) Evaluation of a Korean Lip-sync system for an android robot 2016 13th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI)10.1109/URAI.2016.7734025978-1-5090-0821-6 ACM Transactions on Graphics35:4 (1-11)11 Jul 2016JALIPif Edwards, Chris Landreth, Eugene Fiume and Karan Singh Image and Vision Computing51 (1-12)1 Jul 2016Visual units and confusion modelling for automatic lip-readingDominic Howell, Stephen Cox and Barry Theobald Ibrahim Almajai, Stephen Cox, Richard Harvey and Yuxuan Lan (2016) Improved speaker independent lip reading using speaker adaptive training and deep neural networks 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP.2016.7472172978-1-4799-9988-0 Helen L. Bear and Richard Harvey (2016) Decoding visemes: Improving machine lip-reading 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP.2016.7472029978-1-4799-9988-0 Frontiers in Psychology72 Feb 2016Differential Gaze Patterns on Eyes and Mouth During Audiovisual Speech SegmentationLaina G. Lusk and Aaron D. Mitchel SAGE Open5:4 (215824401561193)23 Dec 2015A Novel Approach for Allocating Mathematical Expressions to Visual Speech SignalsMohammad Hossein Sadaghiani, Niusha Shafiabady and Dino Isa Computational Cognitive Science1:11 Dec 2015Transforming an embodied conversational agent into an efficient talking head: from keyframe-based animation to multimodal concatenation synthesisGuillaume Gibert, Kirk N. Olsen, Yvonne Leung and Catherine J. Stevens Oscar Koller, Hermann Ney and Richard Bowden (2015) Deep Learning of Mouth Shapes for Sign Language 2015 IEEE International Conference on Computer Vision Workshop (ICCVW)10.1109/ICCVW.2015.69978-1-4673-9711-7 Psychonomic Bulletin & Review22:5 (1299-1307)1 Oct 2015Variability and stability in the McGurk effect: contributions of participants, stimuli, time, and response typeDebshila Basu Mallick, John F. Magnotti and Michael S. Beauchamp Cortex68 (169-181)1 Jul 2015Prediction and constraint in audiovisual speech perceptionJonathan E. Peelle and Mitchell S. Sommers Sarah Taylor, Barry-John Theobald and Iain Matthews (2015) A mouth full of words: Visually consistent acoustic redubbing ICASSP 2015 - 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP.2015.7178903978-1-4673-6997-8 Speech Communication66 (182-217)1 Feb 2015Audiovisual speech synthesis: An overview of the state-of-the-artWesley Mattheyses and Werner Verhelst Simon Elias Bibri (2015) Towards AmI Systems Capable of Engaging in 'Intelligent Dialog' and 'Mingling Socially with Humans' The Human Face of Ambient Intelligence10.2991/978-94-6239-130-7_7 PLoS ONE9:12 (e114439)4 Dec 2014Nonnative Audiovisual Speech Perception in Noise: Dissociable Effects of the Speaker and ListenerZilong Xie, Han-Gyol Yi, Bharath Chandrasekaran and Joel Snyder Frontiers in Neuroscience81 Dec 2014Neural pathways for visual speech perceptionLynne E. Bernstein and Einat Liebenthal Communication Disorders Quarterly35:4 (191-203)1 Aug 2014A Review of the Evidence on Strategies for Teaching Children Who Are DHH Grapheme–Phoneme CorrespondenceStacey L. Tucci, Jessica W. Trussell and Susan R. Easterbrooks Frontiers in Psychology516 Jul 2014Talker variability in audio-visual speech perceptionShannon L. M. Heald and Howard C. Nusbaum Lennart Gustafsson, Tamas Jantvik and Andrew P. Paplinski (2014) A Self-organized artificial neural network architecture that generates the McGurk effect 2014 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN.2014.6889411978-1-4799-1484-5 Sarah Taylor, Barry-John Theobald and Iain Matthews (2014) The effect of speaking rate on audio and visual speech ICASSP 2014 - 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP.2014.6854158978-1-4799-2893-4 IEEE Journal of Selected Topics in Signal Processing8:2 (336-347)Joint Audiovisual Hidden Semi-Markov Model-Based Speech SynthesisDietmar Schabus, Michael Pucher and Gregor Hofer Journal of Cognitive Neuroscience26:3 (606-620)1 Mar 2014Audiovisual Speech Integration Does Not Rely on the Motor System: Evidence from Articulatory Suppression, the McGurk Effect, and fMRIWilliam Matchin, Kier Groulx and Gregory Hickok Oscar Koller, Hermann Ney and Richard Bowden (2014) Read My Lips: Continuous Signer Independent Weakly Supervised Viseme Recognition Computer Vision – ECCV 201410.1007/978-3-319-10590-1_19 Mohammad Mahdi Dehshibi, Meysam Alavi and Jamshid Shanbehzadeh (2013) Kernel-based Persian viseme clustering 2013 13th International Conference on Hybrid Intelligent Systems (HIS)10.1109/HIS.2013.6920468978-1-4799-2439-4 Liyan Chen and Beizhan Wang (2013) Research on digital reconstruction of Chinese ancient architecture 2013 International Conference on Anti-Counterfeiting, Security and Identification (ASID)10.1109/ICASID.2013.6825319978-1-4799-1111-0 Speech Communication55:7-8 (857-876)1 Sep 2013Comprehensive many-to-many phoneme-to-viseme mapping and its application for concatenative visual speech synthesisWesley Mattheyses, Lukas Latacz and Werner Verhelst Multimedia Tools and Applications65:3 (521-541)1 Aug 2013Clustering Persian viseme using phoneme subspace for developing visual speech applicationMohammad Aghaahmadi, Mohammad Mahdi Dehshibi, Azam Bastanfard and Mahmood Fazlali Mohammad Hossein Sadaghiani and Fawaz Yahya Annaz (2012) Analysis of visual speech signatures 2012 4th International Congress on Engineering Education (ICEED 2012)10.1109/ICEED.2012.6779263978-1-4673-4868-3 IEEE Transactions on Audio, Speech, and Language Processing20:7 (1936-1947)Language Identification Using Visual FeaturesJacob L. Newman and Stephen J. Cox Yuxuan Lan, Barry-John Theobald and Richard Harvey (2012) View Independent Computer Lip-Reading 2012 IEEE International Conference on Multimedia and Expo (ICME)10.1109/ICME.2012.192978-1-4673-1659-0 Neuropsychologia50:7 (1316-1326)1 Jun 2012Visemic processing in audiovisual discrimination of natural speech: A simultaneous fMRI–EEG studyCyril Dubois, Hélène Otzenberger, Daniel Gounot, Rudolph Sock and Marie-Noëlle Metz-Lutz Journal of Deaf Studies and Deaf Education17:2 (227-243)1 Apr 2012A Dual-Route Cascaded Model of Reading by Deaf Adults: Evidence for Grapheme to Viseme ConversionE. A. Elliott, M. Braun, M. Kuhlmann and A. M. Jacobs Yuxuan Lan, Richard Harvey and Barry-John Theobald (2012) Insights into machine lip reading ICASSP 2012 - 2012 IEEE International Conference on Acoustics, Speech and Signal Processing10.1109/ICASSP.2012.6288999978-1-4673-0046-9 Canadian Journal of Linguistics/Revue canadienne de linguistique57:1 (109-132)1 Mar 2012Anticipation labiale et perception visuelleJohanna-Pascale Roy Nooraini Yusoff, Ioana Sporea and André Grüning (2012) Neural Networks in Cognitive Science Biologically Inspired Networking and Sensing10.4018/978-1-61350-092-7.ch004 Experimental Brain Research215:2 (141-161)1 Nov 2011The perception of visible speech: estimation of speech rate and detection of time reversalsPaolo Viviani, Francesca Figliozzi and Francesco Lacquaniti The Journal of the Acoustical Society of America130:3 (1663-1672)1 Sep 2011Sizing up the competition: Quantifying the influence of the mental lexicon on auditory and visual spoken word recognitionJulia F. Strand and Mitchell S. Sommers Speech Communication53:7 (955-972)1 Sep 2011Evaluation of synthetic and natural Mandarin visual speech: Initial consonants, single vowels, and syllablesTrevor H. Chen and Dominic W. Massaro Pascal Damien (2011) Visual speech recognition of Modern Classic Arabic language 2011 International Symposium on Humanities, Science and Engineering Research (SHUSER)10.1109/SHUSER.2011.6008499978-1-4577-0263-1 O. Gambino, A. Caronia, V. Di Bella, R. Pirrone and S. Gaglio (2011) A M3G Talking Head for Smartphones 2011 International Conference on Complex, Intelligent and Software Intensive Systems (CISIS)10.1109/CISIS.2011.98978-1-61284-709-2 Ophthalmic and Physiological Optics31:3 (249-257)1 May 2011The effect of simulated visual impairment on speech-reading abilityChristine M Dickinson and John Taylor Aphasiology25:2 (202-220)14 Jan 2011Facilitation of picture-naming in anomic subjects: Sound vs mouth shapeAnja Wunderlich and Wolfram Ziegler NeuroReport21:18 (1146-1151)1 Dec 2010Decreased BOLD responses in audiovisual processingEsther Wiersinga-Post, Sonja Tomaskovic, Lavinia Slabu, Remco Renken, Femke de Smit and Hendrikus Duifhuis Ioana Sporea and Andre Gruning (2010) A distributed model of memory for the McGurk effect 2010 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN.2010.5596932978-1-4244-6916-1 Jacob L Newman and Stephen J Cox (2010) Speaker independent visual-only language identification 2010 IEEE International Conference on Acoustics, Speech and Signal Processing10.1109/ICASSP.2010.5495071978-1-4244-4295-9 Yun Fu, Hao Tang, Jilin Tu, Hai Tao and Thomas S. Huang (2010) Human-Centered Face Computing in Multimedia Interaction and Communication Intelligent Multimedia Communication: Techniques and Applications10.1007/978-3-642-11686-5_16 Guanyong Wu, Jie Zhu and Haihua Xu (2009) A hybrid visual feature extraction method for audio-visual speech recognition 2009 16th IEEE International Conference on Image Processing ICIP 200910.1109/ICIP.2009.5413573978-1-4244-5653-6 Scandinavian Journal of Psychology50:5 (419-425)1 Oct 2009Spoken word recognition by eyeEDWARD T. AUER JR Language and Speech52:2-3 (351-367)1 Jun 2009MushyPeek: A Framework for Online Investigation of Audiovisual Dialogue PhenomenaJens Edlund and Jonas Beskow Jacob L Newman and Stephen J Cox (2009) Automatic visual-only language identification: A preliminary study ICASSP 2009 - 2009 IEEE International Conference on Acoustics, Speech and Signal Processing10.1109/ICASSP.2009.4960591978-1-4244-2353-8 PLoS ONE4:3 (e4638)4 Mar 2009Lip-Reading Aids Word Recognition Most in Moderate Noise: A Bayesian Explanation Using High-Dimensional Feature SpaceWei Ji Ma, Xiang Zhou, Lars A. Ross, John J. Foxe, Lucas C. Parra and David Whitney IEEE Transactions on Audio, Speech, and Language Processing17:3 (459-468)Emphatic Visual Speech SynthesisJavier Melenchon, Elisa Martinez, Fernando De La Torre and JosÉ A. Montero Journal of Experimental Child Psychology102:1 (40-59)1 Jan 2009Developmental shifts in children's sensitivity to visual speech: A new multimodal picture–word taskSusan Jerger, Markus F. Damian, Melanie J. Spence, Nancy Tye-Murray and Herve Abdi O. Gambino, A. Augello, A. Caronia, G. Pilato, R. Pirrone and S. Gaglio (2009) A Web-Oriented Java3D Talking Head Human-Computer Systems Interaction10.1007/978-3-642-03202-8_24 Journal of Speech, Language, and Hearing Research51:6 (1607-1619)1 Dec 2008The Effects of Auditory–Visual Vowel Identification Training on Speech Recognition Under Difficult Listening ConditionsCarolyn Richie and Diane Kewley-Port The Journal of the Acoustical Society of America123:1 (428-440)1 Jan 2008Auditory-visual speech perception in normal-hearing and cochlear-implant listenersSheetal Desai, Ginger Stickney and Fan-Gang Zeng IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans37:6 (1063-1076)Review of Low Frame Rate Effects on Human PerformanceJessie Y. C. Chen and Jennifer E. Thropp Perception & Psychophysics69:7 (1070-1083)1 Oct 2007Similarity structure in visual speech perception and optical phonetic signalsJintao Jiang, Edward T. Auer, Abeer Alwan, Patricia A. Keating and Lynne E. Bernstein Antonio Moura, Ingrida Mazonaviciute, Joao Nunes and Justinas Grigaravicius (2007) Human lips synchronisation in Autodesk Maya 2007 14th International Workshop on Systems, Signals and Image Processing and 6th EURASIP Conference focused on Speech and Image Processing, Multimedia Communications and Services10.1109/IWSSIP.2007.4381117978-961-248-036-3 Perception & Psychophysics69:2 (218-231)1 Feb 2007Discriminating languages by speech-readingSalvador Soto-Faraco, Jordi Navarra, Whitney M. Weikum, Athena Vouloumanos, Núria Sebastián-Gallés and Janet F. Werker Computers & Graphics30:6 (971-980)1 Dec 2006Facial animation based on context-dependent visemesJosé Mario De Martino, Léo Pini Magalhães and Fábio Violaro IEEE Transactions on Audio, Speech and Language Processing14:3 (1082-1089)Visual model structures and synchrony constraints for audio-visual speech recognitionT.J. Hazen International Journal of Computer Vision67:3 (297-312)1 May 2006Representation Analysis and Synthesis of Lip Images Using Dimensionality ReductionMichal Aharon and Ron Kimmel Antje Schweitzer, Norbert Braunschweiler, Grzegorz Dogil, Tanja Klankert, Bernd Möbius, Gregor Möhler, Edmilson Morais, Bettina Säuberlich and Matthias Thomae (2006) Multimodal Speech Synthesis SmartKom: Foundations of Multimodal Dialogue Systems10.1007/3-540-36678-4_27 Computer Animation and Virtual Worlds16:3-4 (291-303)1 Jul 2005A feature-based approach to facial expression cloningBongcheol Park, Heejin Chung, Tomoyuki Nishita and Sung Yong Shin Speech Communication46:1 (73-93)1 May 2005Variability in bimodal spoken language processing by native and nonnative speakers of English: A closer look at effects of speech styleDebra M. Hardison Speech Communication44:1-4 (141-154)1 Oct 2004An articulation model for audiovisual speech synthesis—Determination, adjustment, evaluationSascha Fagel and Caroline Clemens Journal of Speech, Language, and Hearing Research46:6 (1367-1377)1 Dec 2003Contributions of Semantic and Facial Information to Perception of Nonsibilant FricativesAllard Jongman, Yue Wang and Brian H. Kim Proceedings of the IEEE91:9 (1406-1429)1 Sep 2003Lifelike talking faces for interactive servicesE. Cosatto, J. Ostermann, H.P. Graf and J. Schoeter Perception32:8 (921-936)1 Aug 2003The Role of Facial Colour and Luminance in Visual and Audiovisual Speech PerceptionMaxine V McCotter and Timothy R Jordan Ear and Hearing23:5 (439-449)1 Oct 2002Auditory-Visual Speech Perception and AgingKathleen M. Cienkowski and Arlene Earley Carney Psychonomic Bulletin & Review9:2 (341-347)1 Jun 2002The influence of the lexicon on speech read word recognition: Contrasting segmental and lexical distinctivenessEdward T. Auer Perception & Psychophysics64:4 (667-679)1 May 2002Stimulus-based lexical distinctiveness as a general word-recognition mechanismSven L. Mattys, Lynne E. Bernstein and Edward T. Auer Jintao Jiang, Abeer Alwan, Lyme E. Bernstein, Edward T. Auer and Patricia A. Keating (2002) Similarity structure in perceptual and physical measures for visual Consonants across talkers Proceedings of ICASSP '0210.1109/ICASSP.2002.57437490-7803-7402-9 Soonkyu Lee and DongSuk Yook (2002) Audio-to-Visual Conversion Using Hidden Markov Models PRICAI 2002: Trends in Artificial Intelligence10.1007/3-540-45683-X_60 Soonkyu Lee and Dongsuk Yook (2002) Viseme Recognition Experiment Using Context Dependent Hidden Markov Models Intelligent Data Engineering and Automated Learning — IDEAL 200210.1007/3-540-45675-9_84 Philosophical Psychology14:3 (261-291)1 Sep 2001Metaphysics, method, and the mouth: Philosophical lessons of speech perceptionJ.D. Trout IEEE Signal Processing Magazine18:1 (9-21)1 Jan 2001Audiovisual speech processing Tsuhan Chen Perception & Psychophysics62:2 (233-252)1 Jan 2000Speech perception without hearingLynne E. Bernstein, Paula E. Tucker and Marilyn E. Demorest C.M. Jones and S.S. Dlay The face as an interface: the new paradigm for HCI IEEE SMC'99 Conference Proceedings. 1999 IEEE International Conference on Systems, Man, and Cybernetics10.1109/ICSMC.1999.8141890-7803-5731-0 Speech Communication26:1-2 (45-63)1 Oct 1998Modeling the interaction of phonemic intelligibility and lexical structure in audiovisual word recognitionPaul Iverson, Lynne E. Bernstein and Edward T. Auer Jr Proceedings of the IEEE86:5 (837-852)1 May 1998Audio-visual integration in multimodal communication Tsuhan Chen and R.R. Rao Journal of Speech, Language, and Hearing Research41:1 (94-106)1 Feb 1998Age-Related Changes on a Children's Test of Sensory-Level Speech Perception CapacityTheresa E. Hnath-Chisolm, Erin Laipply and Arthur Boothroyd T. Ezzat and T. Poggio MikeTalk: a talking facial display based on morphing visemes Computer Animation '9810.1109/CA.1998.6819130-8186-8541-7 J.J. Williams, J.C. Rutledge, D.C. Garstecki and A.K. Katsaggelos Frame rate and viseme analysis for multimedia applications First Signal Processing Society Workshop on Multimedia Signal Processing10.1109/MMSP.1997.6026060-7803-3780-8 IEEE Transactions on Image Processing6:8 (1192-1195)1 Jan 1997Lipreading from color videoG.I. Chiou and Jenq-Neng Hwang Alan J. Goldschen, Oscar N. Garcia and Eric D. Petajan (1997) Continuous Automatic Speech Recognition by Lipreading Motion-Based Recognition10.1007/978-94-015-8935-2_14 Language Learning46:1 (3-73)1 Mar 1996Bimodal Speech Perception by Native and Nonnative Speakers of English: Factors Influencing the McGurk EffectDebra M. Hardison B. Le Goff and C. Benoit A text-to-audiovisual-speech synthesizer for French Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '9610.1109/ICSLP.1996.6072320-7803-3555-4 A.J. Goldschen, O.N. Garcia and E. Petajan Continuous optical automatic speech recognition by lipreading 1994 28th Asilomar Conference on Signals, Systems and Computers10.1109/ACSSC.1994.4715170-8186-6405-3 C. Bregler, S. Manke, H. Hild and A. Waibel Bimodal sensor integration on the example of 'speechreading' IEEE International Conference on Neural Networks10.1109/ICNN.1993.2986340-7803-0999-5 C. Bregler, H. Hild, S. Manke and A. Waibel (1993) Improving connected letter recognition by lipreading Proceedings of ICASSP '9310.1109/ICASSP.1993.3191790-7803-0946-4 D.G. Stork, G. Wolff and E. Levine Neural network lipreading system for improved speech recognition [1992] IJCNN International Joint Conference on Neural Networks10.1109/IJCNN.1992.2269940-7803-0559-0 Language and Speech33:2 (121-135)1 Apr 1990Auditory and Visual Influences on Phonemic RestorationJ.D. Trout and William J. Poser Proceedings of the IEEE78:10 (1658-1668)1 Jan 1990Neural network models of sensory integration for improved vowel recognitionB.P. Yuhas, M.H. Goldstein, T.J. Sejnowski and R.E. Jenkins IEEE Communications Magazine27:11 (65-71)Integration of acoustic and visual speech signals using neural networksB.P. Yuhas, M.H. Goldstein and T.J. Sejnowski Journal of the Optical Society of America A6:4 (491)1 Apr 1989Speechreading with reduced vision: a problem of agingFrank Thorn and Sondra Thorn International Journal of Audiology28:2 (71-81)1 Jan 1989Développement d'une épreuve franco-québécoise de lecture labialeNicole M. Lalande, Ginette Lafleur and Yves S. Lacouture Journal of Phonetics11:4 (323-336)1 Oct 1983Effect of stimulus range on the identification of distorted /r/ in synthesized adult and child /r-w/ continuaDonald J. Sharf and Ralph N. Ohde B.H. Brown, G.S. Dodgson, J.C. Stevens and I.L. Freeston (1983) A wrist-worn, sensory electrical substitution aid for the profoundly deaf High Technology Aids for the Disabled10.1016/B978-0-407-00256-2.50019-8 Scandinavian Audiology8:1 (3-8)1 Jan 1979The Effects of Age Upon the Visual Perception of SpeechCynthia Shoop and C. A. Binnie Australian Journal of Human Communication Disorders5:2 (126-133)1 Dec 1977The Critical Temporal Interval Required for Judgements of Asynchronous Acoustic-Visual Speech SignalsRonald J. Balthazor British Journal of Audiology9:1 (14-18)1 Jan 1975Discrimination of Fricatives by Hearing Impaired Children Using A Vibrotactile CueDaniel Ling and Beverly Sofin Journal of Communication Disorders6:1 (44-52)1 Jan 1973A comparison of visibility and speechreading performance on English and SlurvianPaulette Albright, Nicholas M. Hipskind and Gordon H. Schuckers Volume 11Issue 4December 1968Pages: 796-804 Get Permissions Add to your Mendeley library HistoryReceived: Mar 22, 1968 Published in issue: Dec 1, 1968PubMed ID: 5719234 Metrics Topicsasha-topicsasha-article-typesCopyright & PermissionsCopyright © 1968 American Speech-Language-Hearing AssociationPDF downloadLoading ...

Referência(s)