Recent developments in large language models (LLMs), like those powering AI chatbots such as ChatGPT, have revolutionized text generation by predicting word sequences. However, integrating these models with brain recordings has remained a significant challenge. The key question is whether we can directly generate natural language from brain activity, bypassing the limitations of predefined word sets. A groundbreaking study has made strides toward this vision by introducing BrainLLM, a system that merges brain recordings with LLMs to generate coherent natural language.
In this study, researchers developed BrainLLM, a system designed to directly translate brain activity into natural language by integrating brain recordings with an LLM. The study relied on non-invasive functional magnetic resonance imaging (fMRI) data collected from participants while they processed spoken or written language stimuli. To train the model, the researchers used three public datasets containing fMRI recordings of individuals exposed to various linguistic stimuli.
The key innovation of BrainLLM lies in the creation of a “brain adapter”—a neural network that interprets brain signals and converts them into a format that can be understood by the LLM. This adapter extracted meaningful features from brain activity and combined them with traditional text-based inputs, enabling the LLM to generate words that closely mirrored the linguistic information encoded in brain signals.
The researchers began by collecting brain activity data as participants engaged with language stimuli—either listening to or reading text. This data was then converted into a mathematical representation of brain activity. A specialized neural network mapped these representations into a space compatible with the LLM’s text embeddings. The LLM then processed the combined brain and text inputs, generating word sequences based on both brain activity and existing textual prompts.
Unlike earlier methods that required selecting words from a restricted set, BrainLLM allowed for the generation of continuous text, offering greater flexibility in its output. The system was trained using thousands of brain scans and corresponding language inputs, allowing it to fine-tune its ability to predict words that aligned closely with brain activity.
The team tested BrainLLM on a variety of language tasks, such as predicting the next word in a sequence, reconstructing entire passages, and comparing generated text with human-perceived language continuations. BrainLLM outperformed traditional classification-based models, producing more coherent and contextually appropriate text when processing brain recordings.
One of the study’s key breakthroughs was BrainLLM’s ability to generate open-ended sentences rather than relying on predefined options. This ability marks a major leap toward real-world applications, where free-flowing communication is vital. Human evaluators found that BrainLLM-generated text was more meaningful and linguistically accurate than text produced by baseline models. Notably, the system was particularly adept at reconstructing “surprising” language—words or phrases that would be difficult for an LLM alone to predict—highlighting how brain signals can enhance language modeling in unexpected ways.
The system performed best when analyzing brain activity from regions known for language processing, such as Broca’s area and the auditory cortex. The highest accuracy was seen when using signals from Broca’s area, which is central to speech and language production. This discovery suggests that refining brain signal mapping could improve the accuracy and reliability of language reconstruction in the future.
Despite its impressive performance, BrainLLM still faced challenges. The accuracy of the model varied across individuals, and the system struggled with fully open-ended language reconstruction from brain signals. The study also noted the limitations of fMRI, such as its high cost and complexity, which make it impractical for real-time applications.
The researchers are optimistic about the future of brain-to-text technology. They suggest that electroencephalography (EEG), which is more affordable and capable of real-time brain signal decoding, could offer a more practical alternative to fMRI for brain-computer interfaces. Moreover, integrating BrainLLM with motor-based brain-computer interfaces (BCIs), which have been used to aid communication for individuals with movement disorders, could lead to more robust neuroprosthetic systems.
By further refining brain signal decoding and machine learning algorithms, this research moves us closer to a future where thoughts could be directly translated into words, potentially revolutionizing communication for individuals with speech disabilities and opening the door to unprecedented advancements in brain-computer interfaces.
This study represents a significant step forward in the development of brain-to-text technology. By integrating brain recordings with large language models, BrainLLM has demonstrated that it’s possible to enhance natural language generation using brain activity. Although real-world applications are still years away, this research provides the foundation for future advancements in brain-computer interfaces, with the potential to transform how we communicate.
By Impact Lab

