Visually Guide Motor In Speech
Technology recently allowed researchers to switch from 2D to the Opti-Speech technology, which shows the 3D images of the tongue. Credit: University of Texas at Dallas
A new study done by University of Texas (UT) at Dallas researchers indicates that watching 3D images of tongue movements can help individuals learn speech sounds.
According to Dr. William Katz, co-author of the study and professor at UT Dallas’ Callier Center for Communication Disorders, the findings could be especially helpful for stroke patients seeking to improve their speech articulation.
“These results show that individuals can be taught consonant sounds in part by watching 3D tongue images,” said Katz, who teaches in the UT Dallas School of Behavioral and Brain Sciences. “But we also are seeking to use visual feedback to get at the underlying nature of apraxia and other related disorders.”
The study, which appears in the journal Frontiers in Human Neuroscience, was small but showed that participants became more accurate in learning new sounds when they were exposed to visual feedback training.
Katz is one of the first researchers to suggest that the visual feedback on tongue movements could help stroke patients recover speech.
“People with apraxia of speech can have trouble with this process. They typically know what they want to say but have difficulty getting their speech plans to the muscle system, causing sounds to come out wrong,” Katz said.
“My original inspiration was to show patients their tongues, which would clearly show where sounds should and should not be articulated,” he said.
Technology recently allowed researchers to switch from 2D technology to the Opti-Speech technology, which shows the 3D images of the tongue. A previous UT Dallas research project determined that the Opti-Speech visual feedback system can reliably provide real-time feedback for speech learning.
Part of the new study looked at an effect called compensatory articulation—when acoustics are rapidly shifted and subjects think they are making a certain sound with their mouths, but hear feedback that indicates they are making a different sound.
Katz said people will instantaneously shift away from the direction that the sound has pushed them. Then, if the shift is turned off, they’ll overshoot.
“In our paradigm, we were able to visually shift people. Their tongues were making one sound but, little by little, we start shifting it,” Katz said. “People changed their sounds to match the tongue image.”
Katz said the research results highlight the importance of body visualization as part of rehabilitation therapy, saying there is much more work to be done.
“We want to determine why visual feedback affects speech,” Katz said. “How much is due to compensating, versus mirroring (or entrainment)? Do some of the results come from people visually guiding their tongue to the right place, then having their sense of ‘mouth feel’ take over? What parts of the brain are likely involved?
“3D imaging is opening an entirely new path for speech rehabilitation. Hopefully, this work can be translated soon to help patients who desperately want to speak better.”
Note: Material may have been edited for length and content. For further information, please contact the cited source.
University of Texas at Dallas press release
Katz WF, Mehta S Visual Feedback of Tongue Movement for Novel Speech Sound Learning. Frontiers in Human Neuroscience, Published November 19 2015. doi: 10.3389/fnhum.2015.00612