Scientists have developed AI that can turn brain activity into text. While the system currently works on neural patterns detected while someone is speaking aloud, experts say it could eventually aid communication for patients who are unable to speak or type, such as those with locked-in syndrome.
Makin and colleagues reveal how they developed their system by recruiting four participants who had electrode arrays implanted in their brains to monitor epileptic seizures.
These participants were asked to read aloud from 50 set sentences multiple times, including “Tina Turner is a pop singer”, and “Those thieves stole 30 jewels”. The team tracked their neural activity while they were speaking.
This data was then fed into a machine-learning algorithm, a type of artificial intelligence system that converted the brain activity data for each spoken sentence into a string of numbers.
To make sure the numbers related only to aspects of speech, the system compared sounds predicted from small chunks of the brain activity data with actual recorded audio. The string of numbers was then fed into a second part of the system which converted it into a sequence of words.
At first the system spat out nonsense sentences. But as the system compared each sequence of words with the sentences that were actually read aloud it improved, learning how the string of numbers related to words, and which words tend to follow each other.
The team then tested the system, generating written text just from brain activity during speech.
However, the team found the accuracy of the new system was far higher than previous approaches. While accuracy varied from person to person, for one participant just 3% of each sentence on average needed correcting – higher than the word error rate of 5% for professional human transcribers. But, the team stress, unlike the latter, the algorithm only handles a small number of sentences.
“If you try to go outside the [50 sentences used] decoding gets much worse,” said Makin, adding that the system is likely relying on a combination of learning particular sentences, identifying words from brain activity, and recognizing general patterns in English. The team also found that training the algorithm on one participant’s data meant less training data was needed from the final user – something that could make training less onerous for patients.
Dr. Christian Herff an expert in the field from Maastricht University who was not involved in the study said the research was exciting because the system used less than 40 minutes of training data for each participant, and a limited collection of sentences, rather than the millions of hours typically needed.“By doing so they achieve levels of accuracy that haven’t been achieved so far,” he said.
He noted the system was not yet usable for many severely disabled patients as it relied on the brain activity recorded from people speaking a sentence out loud.“Of course this is fantastic research but those people could just use ‘OK Google’ as well,” he said. “This is not a translation of thought [but of brain activity involved in speech].”
Herff said people should not worry about others reading their thoughts just yet: the brain electrodes must be implanted, while imagined speech is very different from the inner voice.
But Dr. Mahnaz Arvaneh, an expert in brain-machine interfaces at Sheffield University, said it was important to consider ethical issues now. “We [are still] very, very far away from the point that machines can read our minds,” she said. “But it doesn’t mean that we should not think about it and we should not plan about it.”