A research group has built a dataset of the physical movements which create the sounds of speech and could one day be used to develop speech recognition systems that synthesize voices for people with speech impairments.
The database was built using a combination of lip reading and analyzing facial movements. The team used channel impulse response data from radio ultra-wideband and frequency modulated continuous wave radars to capture the movement of the skin of participants’ faces as they spoke, as well as the movements of their tongue and voice box.