Via New Scientist
 
-----
 
 
 
As you read this, your neurons are firing – that brain activity can now be decoded to reveal the silent words in your head
 
TALKING to yourself used to be a strictly private 
pastime. That's no longer the case – researchers have eavesdropped on 
our internal monologue for the first time. The achievement is a step 
towards helping people who cannot physically speak communicate with the 
outside world.
 
"If you're reading text in a newspaper or a book, you hear a voice in your own head," says Brian Pasley
 at the University of California, Berkeley. "We're trying to decode the 
brain activity related to that voice to create a medical prosthesis that
 can allow someone who is paralysed or locked in to speak."
 
When you hear someone speak, sound waves 
activate sensory neurons in your inner ear. These neurons pass 
information to areas of the brain where different aspects of the sound 
are extracted and interpreted as words.
 
In a previous study, Pasley and his 
colleagues recorded brain activity in people who already had electrodes 
implanted in their brain to treat epilepsy, while they listened to 
speech. The team found that certain neurons in the brain's temporal lobe
 were only active in response to certain aspects of sound, such as a 
specific frequency. One set of neurons might only react to sound waves 
that had a frequency of 1000 hertz, for example, while another set only 
cares about those at 2000 hertz. Armed with this knowledge, the team 
built an algorithm that could decode the words heard based on neural activity alone (PLoS Biology, doi.org/fzv269).
 (PLoS Biology, doi.org/fzv269).
 
The team hypothesised that hearing speech 
and thinking to oneself might spark some of the same neural signatures 
in the brain. They supposed that an algorithm trained to identify speech
 heard out loud might also be able to identify words that are thought.
 
 
 
Mind-reading
 
To test the idea, they recorded brain 
activity in another seven people undergoing epilepsy surgery, while they
 looked at a screen that displayed text from either the Gettysburg Address, John F. Kennedy's inaugural address or the nursery rhyme Humpty Dumpty.
 
Each participant was asked to read the 
text aloud, read it silently in their head and then do nothing. While 
they read the text out loud, the team worked out which neurons were 
reacting to what aspects of speech and generated a personalised decoder 
to interpret this information. The decoder was used to create a 
spectrogram – a visual representation of the different frequencies of 
sound waves heard over time. As each frequency correlates to specific 
sounds in each word spoken, the spectrogram can be used to recreate what
 had been said. They then applied the decoder to the brain activity that
 occurred while the participants read the passages silently to 
themselves (see diagram).
 
Despite the neural activity from imagined 
or actual speech differing slightly, the decoder was able to reconstruct
 which words several of the volunteers were thinking, using neural 
activity alone (Frontiers in Neuroengineering, doi.org/whb).
 
The algorithm isn't perfect, says 
Stephanie Martin, who worked on the study with Pasley. "We got 
significant results but it's not good enough yet to build a device."
 
 
In
 practice, if the decoder is to be used by people who are unable to 
speak it would have to be trained on what they hear rather than their 
own speech. "We don't think it would be an issue to train the decoder on
 heard speech because they share overlapping brain areas," says Martin.
 
The team is now fine-tuning their 
algorithms, by looking at the neural activity associated with speaking 
rate and different pronunciations of the same word, for example. "The 
bar is very high," says Pasley. "Its preliminary data, and we're still 
working on making it better."
 
The team have also turned their hand to 
predicting what songs a person is listening to by playing lots of Pink 
Floyd to volunteers, and then working out which neurons respond to what 
aspects of the music. "Sound is sound," says Pasley. "It all helps us 
understand different aspects of how the brain processes it."
 
"Ultimately, if we understand covert 
speech well enough, we'll be able to create a medical prosthesis that 
could help someone who is paralysed, or locked in and can't speak," he 
says.
 
Several other researchers are also investigating ways to read the human mind. Some can tell what pictures a person is looking at, others have worked out what neural activity represents certain concepts in the brain, and one team has even produced crude reproductions of movie clips
 that someone is watching just by analysing their brain activity. So is 
it possible to put it all together to create one multisensory 
mind-reading device?
 
In theory, yes, says Martin, but it would 
be extraordinarily complicated. She says you would need a huge amount of
 data for each thing you are trying to predict. "It would be really 
interesting to look into. It would allow us to predict what people are 
doing or thinking," she says. "But we need individual decoders that work
 really well before combining different senses."