Post

vittorio
vittorio@IterIntellectus·
🚨IT'S HAPPENING🚨 a woman who cannot speak now speaks through her brain, in real time, with her own voice. no typing, delay, or sounds made. just neural intent to streaming speech this isn’t prediction. it’s embodiment 1/
vittorio tweet media
English
478
3.1K
25K
2.4M
vittorio
vittorio@IterIntellectus·
a woman with complete paralysis and zero vocal function just regained the ability to speak at ~conversational speed. with AI, a brain implant learned to read her intention to speak and synthesized her words in her own voice. live. streaming. no delay 2/
English
24
270
2.8K
175.2K
vittorio
vittorio@IterIntellectus·
not typing or text. actual speech. before this, the best systems let patients “type” at 8–14 words per minute. this one does 90+ WPM. in audio, with prosody. from brain activity alone. and it uses no audible training data. she doesn’t even need to try making sounds. 3/
English
7
30
1.3K
111.3K
vittorio
vittorio@IterIntellectus·
the system: - a 253-channel ECoG array on her speech motor cortex - a deep neural decoder trained on 23,000 silent-speech attempts - RNN-T architecture decoding in 80ms chunks - dual output: synthesized speech + real-time transcript - voice personalized to her own, pre-injury
English
5
30
966
102.5K
vittorio
vittorio@IterIntellectus·
the decoder streams speech every 80ms most systems wait for full sentences before outputting anything. this one doesn’t. it emits speech as the brain thinks. lag: ~1 second. the system literally streams her neural intention into speech in near real time. 5/
English
2
18
736
91.1K
vittorio
vittorio@IterIntellectus·
the speech is fast, fluent, and accurate on testing: 50-phrase set (caregiver needs): 91 WPM 12% word error rate (WER) 11% character error rate large 1,024-word set (natural sentences): 47 WPM 59% WER (harder) 45% character error rate it’s not perfect. but it works. 6/
vittorio tweet media
English
3
17
685
82.5K
vittorio
vittorio@IterIntellectus·
it even decoded words it never saw the model successfully synthesized novel words not seen during training. when given 24 new silent words like “Zulu” or “Romeo,” it correctly identified them 46% of the time vs 3.8% by chance. from neural activity alone 7/
vittorio tweet media
English
3
17
598
70.2K
vittorio
vittorio@IterIntellectus·
it also generalizes across tech the same neural architecture decoded speech from: - ECoG (surface brain electrodes) - MEA (intracortical microelectrodes) - EMG (surface facial electrodes, no surgery) silent speech, multiple inputs. one decoder framework. 8/
vittorio tweet media
English
3
16
471
63.2K
vittorio
vittorio@IterIntellectus·
works continuously, not just in trials the system doesn’t need pre-programmed trials or cues it can detect when she starts and stops speaking from brain activity they tested it on 6-minute continuous silent speech blocks. it decoded accurately, with almost no false positives. 9/
vittorio tweet media
English
1
19
449
57.4K
vittorio
vittorio@IterIntellectus·
and no, it’s not just parroting audio is it reacting to auditory feedback? no. they ran trials with and without speech feedback. result: no performance change, decoder works based on motor intent, not sound. it’s intention → speech. 10/
English
5
17
482
52.9K
vittorio
vittorio@IterIntellectus·
this is the new benchmark previous best decoding speed: 28 WPM this system: 90 WPM, at lower latency no need for vocalization and works with generalizable architecture and continuous, volitional use we’re entering the post-vocal age. 11/
English
1
14
445
50.8K
vittorio
vittorio@IterIntellectus·
clinically this gives speech back to people who’ve lost it technically it solves real-time, fluent neural speech decoding this will enable silent communication and shows that language can be purely neural 12/
English
4
21
492
49K
vittorio
vittorio@IterIntellectus·
today: ECoG in paralyzed patients. next: non-invasive silent speech interfaces (EMG → wearable). eventually: ambient, invisible BCI that lets you speak mind-to-mind. and one day we'll have internal language externalized without sound or typing. though to action models 13/
English
4
15
447
47.9K
vittorio
vittorio@IterIntellectus·
imagine 20 years from now: - no phones - no keyboards - no voice commands - you just think and your words are heard not as text or buttons. as your voice, in real time. pure thought to pure speech and action 14/
English
45
28
575
50.5K
Styx
Styx@Colossus_of_·
@IterIntellectus Does it work on people without an inner monologue?
English
1
0
5
146
Paylaş