Whole-word recognition from articulatory movements for silent speech interfaces

Jun Wang, Ashok K Samal, Jordan R. Green, Frank Rudzicz

Research output: Chapter in Book/Report/Conference proceedingConference contribution

14 Citations (Scopus)

Abstract

Articulation-based silent speech interfaces convert silently produced speech movements into audible words. These systems are still in their experimental stages, but have significant potential for facilitating oral communication in persons with laryngectomy or speech impairments. In this paper, we report the result of a novel, real-time algorithm that recognizes wholewords based on articulatory movements. This approach differs from prior work that has focused primarily on phoneme-level recognition based on articulatory features. On average, our algorithm missed 1.93 words in a sequence of twenty-five words with an average latency of 0.79 seconds for each word prediction using a data set of 5,500 isolated word samples collected from ten speakers. The results demonstrate the effectiveness of our approach and its potential for building a real-time articulationbased silent speech interface for health applications.

Original languageEnglish (US)
Title of host publication13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012
Pages1326-1329
Number of pages4
StatePublished - Dec 1 2012
Event13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012 - Portland, OR, United States
Duration: Sep 9 2012Sep 13 2012

Publication series

Name13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012
Volume2

Conference

Conference13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012
CountryUnited States
CityPortland, OR
Period9/9/129/13/12

Fingerprint

human being
Health
communication
health
Communication
time

Keywords

  • Laryngectomy
  • Silent speech recognition
  • Speech impairment
  • Support vector machine

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Communication

Cite this

Wang, J., Samal, A. K., Green, J. R., & Rudzicz, F. (2012). Whole-word recognition from articulatory movements for silent speech interfaces. In 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012 (pp. 1326-1329). (13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012; Vol. 2).

Whole-word recognition from articulatory movements for silent speech interfaces. / Wang, Jun; Samal, Ashok K; Green, Jordan R.; Rudzicz, Frank.

13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012. 2012. p. 1326-1329 (13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012; Vol. 2).

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Wang, J, Samal, AK, Green, JR & Rudzicz, F 2012, Whole-word recognition from articulatory movements for silent speech interfaces. in 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012. 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012, vol. 2, pp. 1326-1329, 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012, Portland, OR, United States, 9/9/12.
Wang J, Samal AK, Green JR, Rudzicz F. Whole-word recognition from articulatory movements for silent speech interfaces. In 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012. 2012. p. 1326-1329. (13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012).
Wang, Jun ; Samal, Ashok K ; Green, Jordan R. ; Rudzicz, Frank. / Whole-word recognition from articulatory movements for silent speech interfaces. 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012. 2012. pp. 1326-1329 (13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012).
@inproceedings{3cd222dc83944bb7ab3e30ce414dc311,
title = "Whole-word recognition from articulatory movements for silent speech interfaces",
abstract = "Articulation-based silent speech interfaces convert silently produced speech movements into audible words. These systems are still in their experimental stages, but have significant potential for facilitating oral communication in persons with laryngectomy or speech impairments. In this paper, we report the result of a novel, real-time algorithm that recognizes wholewords based on articulatory movements. This approach differs from prior work that has focused primarily on phoneme-level recognition based on articulatory features. On average, our algorithm missed 1.93 words in a sequence of twenty-five words with an average latency of 0.79 seconds for each word prediction using a data set of 5,500 isolated word samples collected from ten speakers. The results demonstrate the effectiveness of our approach and its potential for building a real-time articulationbased silent speech interface for health applications.",
keywords = "Laryngectomy, Silent speech recognition, Speech impairment, Support vector machine",
author = "Jun Wang and Samal, {Ashok K} and Green, {Jordan R.} and Frank Rudzicz",
year = "2012",
month = "12",
day = "1",
language = "English (US)",
isbn = "9781622767595",
series = "13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012",
pages = "1326--1329",
booktitle = "13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012",

}

TY - GEN

T1 - Whole-word recognition from articulatory movements for silent speech interfaces

AU - Wang, Jun

AU - Samal, Ashok K

AU - Green, Jordan R.

AU - Rudzicz, Frank

PY - 2012/12/1

Y1 - 2012/12/1

N2 - Articulation-based silent speech interfaces convert silently produced speech movements into audible words. These systems are still in their experimental stages, but have significant potential for facilitating oral communication in persons with laryngectomy or speech impairments. In this paper, we report the result of a novel, real-time algorithm that recognizes wholewords based on articulatory movements. This approach differs from prior work that has focused primarily on phoneme-level recognition based on articulatory features. On average, our algorithm missed 1.93 words in a sequence of twenty-five words with an average latency of 0.79 seconds for each word prediction using a data set of 5,500 isolated word samples collected from ten speakers. The results demonstrate the effectiveness of our approach and its potential for building a real-time articulationbased silent speech interface for health applications.

AB - Articulation-based silent speech interfaces convert silently produced speech movements into audible words. These systems are still in their experimental stages, but have significant potential for facilitating oral communication in persons with laryngectomy or speech impairments. In this paper, we report the result of a novel, real-time algorithm that recognizes wholewords based on articulatory movements. This approach differs from prior work that has focused primarily on phoneme-level recognition based on articulatory features. On average, our algorithm missed 1.93 words in a sequence of twenty-five words with an average latency of 0.79 seconds for each word prediction using a data set of 5,500 isolated word samples collected from ten speakers. The results demonstrate the effectiveness of our approach and its potential for building a real-time articulationbased silent speech interface for health applications.

KW - Laryngectomy

KW - Silent speech recognition

KW - Speech impairment

KW - Support vector machine

UR - http://www.scopus.com/inward/record.url?scp=84878410082&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84878410082&partnerID=8YFLogxK

M3 - Conference contribution

AN - SCOPUS:84878410082

SN - 9781622767595

T3 - 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012

SP - 1326

EP - 1329

BT - 13th Annual Conference of the International Speech Communication Association 2012, INTERSPEECH 2012

ER -