Biosignal-Based Spoken Communication: A Survey
Open Access
- 23 November 2017
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE/ACM Transactions on Audio, Speech, and Language Processing
- Vol. 25 (12), 2257-2271
- https://doi.org/10.1109/taslp.2017.2752365
Abstract
Speech is a complex process involving a wide range of biosignals, including but not limited to acoustics. These biosignals—stemming from the articulators, the articulator muscle activities, the neural pathways, and the brain itself—can be used to circumvent limitations of conventional speech processing in particular, and to gain insights into the process of speech production in general. Research on biosignal-based speech processing is a wide and very active field at the intersection of various disciplines, ranging from engineering, computer science, electronics and machine learning to medicine, neuroscience, physiology, and psychology. Consequently, a variety of methods and approaches have been used to investigate the common goal of creating biosignal-based speech processing devices for communication applications in everyday situations and for speech rehabilitation, as well as gaining a deeper understanding of spoken communication. This paper gives an overview of the various modalities, research approaches, and objectives for biosignal-based spoken communication.Keywords
Funding Information
- Federal Ministry of Education and Research (BMBF)
- National Science Foundation (NSF)
- RESPONSE - REvealing SPONtaneous Speech processes in Electrocorticography
- EU H2020 (#687795)
- National Institutes of Health (R03-DC011304)
This publication has 106 references indexed in Scilit:
- A review and synthesis of the first 20years of PET and fMRI studies of heard speech, spoken language and readingNeuroImage, 2012
- Decoding spoken words using local field potentials recorded from the cortical surfaceJournal of Neural Engineering, 2010
- Brain–computer interfaces for speech communicationSpeech Communication, 2010
- A Wireless Brain-Machine Interface for Real-Time Speech SynthesisPLOS ONE, 2009
- Single-trial classification of vowel speech imagery using common spatial patternsNeural Networks, 2009
- "Who" Is Saying "What"? Brain-Based Decoding of Human Voice and SpeechScience, 2008
- Neural mechanisms underlying auditory feedback control of speechNeuroImage, 2008
- The Utah Intracortical Electrode Array: A recording structure for potential brain-computer interfacesElectroencephalography and Clinical Neurophysiology, 1997
- A tutorial on hidden Markov models and selected applications in speech recognitionProceedings of the IEEE, 1989
- Hearing lips and seeing voicesNature, 1976