Cybermind AI Brain Technology Illustration

Not science fiction: Mind exercise decoder turns ideas into textual content

Researchers on the College of Texas at Austin have developed a semantic decoder that converts mind exercise right into a steady stream of textual content, in keeping with a research printed in Nature Neuroscience. This non-invasive synthetic intelligence system relies on a transformer mannequin and will probably assist individuals who can not bodily talk as a consequence of circumstances equivalent to strokes. Contributors prepare with the decoder by listening to hours of podcasts whereas in an fMRI scanner. The decoder then creates textual content from mind exercise whereas the participant listens or imagines a narrative.

A semantic decoder that converts mind exercise into textual content has been developed by researchers on the College of Texas at Austin. This synthetic intelligence system, which is non-invasive and doesn’t require surgical implants, may present a brand new technique of communication for individuals who can not bodily communicate. The decoder is educated by having the participant hearken to hours of podcasts whereas in a single

fMRI
fMRI stands for practical magnetic resonance imaging. It’s a non-invasive neuroimaging approach that makes use of magnetic resonance imaging (MRI) to measure modifications in blood circulation within the mind, which may point out neural exercise. In easier phrases, fMRI is a device that enables researchers to see which components of the mind are lively throughout particular duties or stimuli, offering details about mind operate and group.

“data-gt-translate-attributes=”[{” attribute=””>fMRI scanner, and it can then generate text based on brain activity alone.

A new artificial intelligence system called a semantic decoder can translate a persons brain activity while listening to a story or silently imagining telling a story into a continuous stream of text. The system developed by researchers at The University of Texas at Austin might help people who are mentally conscious yet unable to physically speak, such as those debilitated by strokes, to communicate intelligibly again.

The study,published today (May 1) in the journal Nature Neuroscience, was led by Jerry Tang, a doctoral student in computer science, and Alex Huth, an assistant professor of neuroscience and computer science at UT Austin. The work relies in part on a transformer model, similar to the ones that power Open AIs ChatGPT and Googles Bard.

Unlike other language decoding systems in development, this system does not require subjects to have surgical implants, making the process noninvasive. Participants also do not need to use only words from a prescribed list. Brain activity is measured using an fMRI scanner after extensive training of the decoder, in which the individual listens to hours of podcasts in the scanner. Later, provided that the participant is open to having their thoughts decoded, their listening to a new story or imagining telling a story allows the machine to generate corresponding text from brain activity alone.

Scientists Peer Through fMRI Scanner

Researchers Alex Huth (left), Jerry Tang (right) and Shailee Jain (center) prepare to collect brain activity data in the Biomedical Imaging Center at The University of Texas at Austin. The researchers trained their semantic decoder on dozens of hours of brain activity data from members of the lab, collected in an fMRI scanner. Credit: Nolan Zunk/University of Texas at Austin

For a noninvasive method, this is a real leap forward compared to whats been done before, which is typically single words or short sentences, Huth said. Were getting the model to decode continuous language for extended periods of time with complicated ideas.

The result is not a word-for-word transcript. Instead, researchers designed it to capture the gist of what is being said or thought, albeit imperfectly. About half the time, when the decoder has been trained to monitor a participants brain activity, the machine produces text that closely (and sometimes precisely) matches the intended meanings of the original words.

For example, in experiments, a participant listening to a speaker say, I dont have my drivers license yet had their thoughts translated as, She has not even started to learn to drive yet. Listening to the words, I didnt know whether to scream, cry or run away. Instead, I said, Leave me alone! was decoded as, Started to scream and cry, and then she just said, I told you to leave me alone.

Text Person Heard Versus Semantic Decoder Comparison

This image shows decoder predictions from brain recordings collected while a user listened to four stories. Example segments were manually selected and annotated to demonstrate typical decoder behaviors. The decoder exactly reproduces some words and phrases and captures the gist of many more. Credit: University of Texas at Austin

Beginning with an earlier version of the paper that appeared as a preprint online, the researchers addressed questions about potential misuse of the technology. The paper describes how decoding worked only with cooperative participants who had participated willingly in training the decoder. Results for individuals on whom the decoder had not been trained were unintelligible, and if participants on whom the decoder had been trained later put up resistance for example, by thinking other thoughts results were similarly unusable.

We take very seriously the concerns that it could be used for bad purposes and have worked to avoid that, Tang said. We want to make sure people only use these types of technologies when they want to and that it helps them.

In addition to having participants listen or think about stories, the researchers asked subjects to watch four short, silent videos while in the scanner. The semantic decoder was able to use their brain activity to accurately describe certain events from the videos.

The system currently is not practical for use outside of the laboratory because of its reliance on the time need on an fMRI machine. But the researchers think this work could transfer to other, more portable brain-imaging systems, such as functional near-infrared spectroscopy (fNIRS).

fNIRS measures where theres more or less blood flow in the brain at different points in time, which, it turns out, is exactly the same kind of signal that fMRI is measuring, Huth said. So, our exact kind of approach should translate to fNIRS, although, he noted, the resolution with fNIRS would be lower.

This work was supported by the Whitehall Foundation, the Alfred P. Sloan Foundation and the Burroughs Wellcome Fund.

The studys other co-authors are Amanda LeBel, a former research assistant in the Huth lab, and Shailee Jain, a computer science graduate student at UT Austin.

Alexander Huth and Jerry Tang have filed a PCT patent application related to this work.

Frequently Asked Questions

Could this technology be used on someone without them knowing, say by an authoritarian regime interrogating political prisoners or an employer spying on employees?

No. The system has to be extensively trained on a willing subject in a facility with large, expensive equipment. A person needs to spend up to 15 hours lying in an MRI scanner, being perfectly still, and paying good attention to stories that theyre listening to before this really works well on them, said Huth.

Could training be skipped altogether?

No. The researchers tested the system on people whom it hadnt been trained on and found that the results were unintelligible.

Are there ways someone can defend against having their thoughts decoded?

Yes. The researchers tested whether a person who had previously participated in training could actively resist subsequent attempts at brain decoding. Tactics like thinking of animals or quietly imagining telling their own story let participants easily and completely thwart the system from recovering the speech the person was exposed to.

What if technology and related research evolved to one day overcome these obstacles or defenses?

I think right now, while the technology is in such an early state, its important to be proactive by enacting policies that protect people and their privacy, Tang said. Regulating what these devices can be used for is also very important.

Reference: Semantic reconstruction of continuous language from non-invasive brain recordings 1 May 2023, Nature Neuroscience.
DOI: 10.1038/s41593-023-01304-9

Leave a Reply

Your email address will not be published. Required fields are marked *