Scientists at Pohang College of Science and Know-how (POSTECH), in South Korea, have constructed a silicone neckband that reads the tiny actions of your neck as you mouth phrases – and turns them into speech in your personal voice, transmitted to whoever is listening.
The system relies on the truth that speech would not solely produce sound. Each time you kind a phrase, your neck muscle groups and pores and skin shift in methods which are delicate however completely predictable – a sort of silent fingerprint for every syllable. Most earlier makes an attempt to seize that sign relied on electromyography (EMG), which measures electrical exercise in muscle groups, or electroencephalography (EEG), which reads brainwaves. Each approaches have dragged the identical baggage for years: cumbersome tools, uncomfortable adhesive electrodes, and efficiency that tended to crumble the second a affected person left the lab.
The POSTECH crew took a distinct route. The neckband combines comfortable silicone, a miniature digital camera, and movement sensors with an AI mannequin educated on the wearer’s personal voice. This Multiaxial Pressure Mapping Sensor tracks not simply how a lot the pores and skin deforms while you communicate but additionally through which route it strikes, a distinction that provides it a far richer image of what your mouth and throat are doing. Reference markers printed instantly onto the silicone collar let a tiny onboard digital camera measure these deformations in actual time.
An algorithm then corrects for the slight positional variations that happen every time the choker is placed on, so it reads persistently even in the event you do not clip it in precisely the identical spot twice. These deformation patterns feed into an AI mannequin that identifies which phrase was articulated.
POSTECH
In checks, the system was educated on the NATO phonetic alphabet (“Alpha,” “Bravo,” “Charlie” and the remaining), a vocabulary particularly designed for intelligibility below troublesome situations. Throughout 26 phrases, it achieved 85.8% accuracy.
As soon as the AI acknowledges a phrase, it sends the end result wirelessly to a server, which synthesizes it as audio utilizing a text-to-speech mannequin personalised to the neckband’s wearer. Researchers say that coaching the voice mannequin requires lower than 10 minutes of recordings, after which the system reproduces the consumer’s personal intonation and vocal character with waveforms they describe as carefully matching the true factor.
The collar additionally held up towards critical background noise. In checks with white noise at roughly 90 dB – roughly the quantity of a busy development website – the system maintained a signal-to-noise ratio of as much as 33.75 dB, which the crew says outperforms industrial EMG methods below the identical situations.
“We hope this expertise will speed up the day when sufferers with speech problems can reclaim their voices,” stated Professor Sung-Min Park, who led the analysis. “It’s a noteworthy expertise as a result of it has a variety of potential purposes, together with aiding laryngectomized sufferers, speaking in noisy industrial environments, and even supporting silent conversations.”
Past medication, the purposes prolong to any setting the place standard microphones fail or are merely not an possibility. The authors of a paper revealed in Cyborg and Bionic Systems particularly level to industrial amenities, emergency response, aviation, maritime operations, and army situations – they usually have put that declare to the take a look at not simply with white noise but additionally throughout a gasoline blowback rifle demonstration, the place each noise and bodily vibration have been in play.
For all its promise, the system nonetheless has important limitations the authors themselves are candid about. It really works solely with a hard and fast vocabulary of 26 predefined phrases – not free dialog – and accuracy can drop to 39.72% when the consumer walks or makes pronounced head actions. The crew’s subsequent steps are testing with extra customers throughout extra periods, increasing the vocabulary, and bettering compensation for physique motion.
This isn’t the primary time we have seen this type of strategy examined within the lab. A few years in the past, researchers on the College of Cambridge additionally went with a sensor-packed choker to detect throat vibrations as a consumer silently mouthed phrases. Their prototype system was reported to realize a speech decoding accuracy of 95.25%, and it seems to be like take a look at topics weren’t restricted to particular phrases.
The Cambridge lab constructed on this work for a follow up earlier this year, which not solely decoded silent speech but additionally detected the emotional state of the wearer. The POSTECH crew will get bonus factors for utilizing AI to approximate the consumer’s personal voice although.
Supply: POSTECH

