The human brain may work more like AI than anyone expected
Scientists have discovered that the human brain understands spoken language in a way that closely resembles how advanced AI language models work. By tracking brain activity as people listened to a long podcast, researchers found that meaning unfolds step by step—much like the layered processing inside systems such as GPT-style models.
A new study suggests that the human brain understands spoken language through a stepwise process that closely resembles how advanced AI language models operate. By recording brain activity from people listening to a spoken story, researchers found that later stages of brain responses match deeper layers of AI systems, especially in well known language regions like Broca's area. The results call into question long standing rule-based ideas of language comprehension and are supported by a newly released public dataset that offers a powerful new way to study how meaning is formed in the brain.
The research, published in Nature Communications, was led by Dr. Ariel Goldstein of the Hebrew University with collaborators Dr. Mariano Schain of Google Research and Prof Uri Hasson and Eric Ham from Princeton University. Together, the team uncovered an unexpected similarity between how humans make sense of speech and how modern AI models process text.
Using electrocorticography recordings from participants who listened to a thirty-minute podcast, the scientists tracked the timing and location of brain activity as language was processed. They found that the brain follows a structured sequence that closely matches the layered design of large language models such as GPT-2 and Llama 2.
How the Brain Builds Meaning Over Time
As we listen to someone speak, the brain does not grasp meaning all at once. Instead, each word passes through a series of neural steps. Goldstein and his colleagues showed that these steps unfold over time in a way that mirrors how AI models handle language. Early layers in AI focus on basic word features, while deeper layers combine context, tone, and broader meaning.
Human brain activity followed the same pattern. Early neural signals matched the early stages of AI processing, while later brain responses lined up with the deeper layers of the models. This timing match was especially strong in higher level language areas such as Broca's area, where responses peaked later when linked to deeper AI layers.
According to Dr. Goldstein, "What surprised us most was how closely the brain's temporal unfolding of meaning matches the sequence of transformations inside large language models. Even though these systems are built very differently, both seem to converge on a similar step-by-step buildup toward understanding"
Why These Findings Matter
The study suggests that artificial intelligence can do more than generate text. It may also help scientists better understand how the human brain creates meaning. For many years, language was thought to rely mainly on fixed symbols and rigid hierarchies. These results challenge that view and instead point to a more flexible and statistical process in which meaning gradually emerges through context.
The researchers also tested traditional linguistic elements such as phonemes and morphemes. These classic features did not explain real time brain activity as well as the contextual representations produced by AI models. This supports the idea that the brain relies more on flowing context than on strict linguistic building blocks.
A New Resource for Language Neuroscience
To help move the field forward, the team has made the complete set of neural recordings and language features publicly available. This open dataset allows researchers around the world to compare theories of language understanding and to develop computational models that more closely reflect how the human mind works.
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellanguage modelresearchExclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxNZUVvVzNvMVZBN09Hc3c5QWoxMUN4MEo1dDIzQjFXd1pkUlVWT25ZQ3pjOS1RMGdONzNrLUpfMHdLYTZUVm5YQkZXSDZxLU5qSkpCR0pYTTRURWtOR2JhMWMtMWVPX1dIRmlwN0lGTkFtVlVaRHdIeEFabFBTQU9FeWFiY3B1NERUNTc1X2N1djhGUk0yYVdUUjFzdGd0eFVuWVZ1TzN1akZCQmtuS3RzNWg3YVNraTFWd0ozX2hISTlUbElnQ29vUUx0WThlUVppU3E5LTNTSllDcXV0dUJUU29mWkVWZDV2OEtRbC1mRWYtWGZQOXUxZ1ppS1B6dThsSXM3V3lfNUxseERsUFVISW1HUXNUSVRhbUNLZ09sSjhDRnNGZXQtdGVsbGZCZXppRUdsdUhrWERta1V5Vm5GeWR6V1pvWE91TUFqQ0tfdU04TkpNTnIwSnd6RGNLWUo2ZVc1dHlzV1lWbi0xdWYtWXJCZ3RwTkdmS1R3Sk5RTl9iTFpTNXVFQktzYWZqYmNRZUpJMHBnRkNUb0tWMG1fV2JzckVhS1Zla0hoOE9Ra2hkdmZFT2lXWA?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>

Polysemanticity or Polysemy? Lexical Identity Confounds Superposition Metrics
arXiv:2604.00443v1 Announce Type: new Abstract: If the same neuron activates for both "lender" and "riverside," standard metrics attribute the overlap to superposition--the neuron must be compressing two unrelated concepts. This work explores how much of the overlap is due a lexical confound: neurons fire for a shared word form (such as "bank") rather than for two compressed concepts. A 2x2 factorial decomposition reveals that the lexical-only condition (same word, different meaning) consistently exceeds the semantic-only condition (different word, same meaning) across models spanning 110M-70B parameters. The confound carries into sparse autoencoders (18-36% of features blend senses), sits in <=1% of activation dimensions, and hurts downstream tasks: filtering it out improves word sense di

TR-ICRL: Test-Time Rethinking for In-Context Reinforcement Learning
arXiv:2604.00438v1 Announce Type: new Abstract: In-Context Reinforcement Learning (ICRL) enables Large Language Models (LLMs) to learn online from external rewards directly within the context window. However, a central challenge in ICRL is reward estimation, as models typically lack access to ground-truths during inference. To address this limitation, we propose Test-Time Rethinking for In-Context Reinforcement Learning (TR-ICRL), a novel ICRL framework designed for both reasoning and knowledge-intensive tasks. TR-ICRL operates by first retrieving the most relevant instances from an unlabeled evaluation set for a given query. During each ICRL iteration, LLM generates a set of candidate answers for every retrieved instance. Next, a pseudo-label is derived from this set through majority voti
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
ChatGPT vs. Claude: 7 real-life benchmarks that crown the 2026 AI Madness Champion - Tom's Guide
<a href="https://news.google.com/rss/articles/CBMirgFBVV95cUxOdUdicEhTLURxU0paMjZJNGNpZzFLVWhmUjZPVEJxazgzR0JsNDFSRUQwZUVIenVWWmo0Z0FJUGMxT0lrb1pyTnlVSDFzcUtmVWFpTHNKY2hUN2p5WWt3UFBxRFJ6bEIwWm9iSjVxMldQRE1HclM2WWtzUk5LMFBmLWRrWkZXcWVzS3N3M3h3d3Y1eWpsdFd4dG5jaGZKektFRXh4dHdWNVRhZDRMR1E?oc=5" target="_blank">ChatGPT vs. Claude: 7 real-life benchmarks that crown the 2026 AI Madness Champion</a> <font color="#6f6f6f">Tom's Guide</font>
Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT - WSJ
<a href="https://news.google.com/rss/articles/CBMiogNBVV95cUxOZWxXTEVKWE9LdjlmTUtCRGt3YTc5aU51d2JfakIzdzdLUHlDNDYwLXl6WWlnMWhFdHM0Rm1ZeDVvRWo4MWEwN19TQTFmQXVnb285bFlLQ18zOTBsMFZOOTZrNkpzNHgya3gwRzRTSjNEWTd5aXd0RkZoMEVZa2xIeGhhRzBhZC02UG5NXzdlX0JUMXIzV3lBRm5FNExoclV4d2ZUY3pCWnoxOFczVDFuWEY3cFZFV2hQRzZhdU05bjZ0b252dHJfSnltUTZ6Z2czZ3RMUm1qQWI4R3lvdUVmOU9IREU4b2lvRi1aM1JhMEsxamtkMjBETkh6YzFrVDFmckQ0eENMNW5NVXNtWXNNNmYzSUE2TTY3T3RzTy1FNEFESWxwUTh2NmpDaFVXczIzZ3RXVk5CTjR4S0g1MllkWjdsN2VGREx0ZWdsZ1AyeFJPd20tSGdEZWhKR2c5alMzTDBrVVZheTZGRjRvRkFRV1duejdIMHhlRFFPLXY2UmIycHZaWExIN3J6djlOcy1KanRwSzg4NkJkaVBQd2MwemNn?oc=5" target="_blank">Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT</a> <font color="#6f6f6f">WSJ</font>
Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT - WSJ
<a href="https://news.google.com/rss/articles/CBMiogNBVV95cUxNUDBsOElrTS1fVG1zWTJYQnZPNXBLY3lLUUZYLVhrUHlmcUI2ak1fTWlKTTFzSy1jSWZFNk5kYnhpdmxyWWxFRnd0TmtJSnE2QXktRVFpRUI2U2pTdThTaDNwY3Rjcm15bDM2ZXpUSm1DOGNwY0dsSjhjUHpHSVhNZ002SndLbDBTa3Q3bDhIazZXTFUwMDVFbzl1R0ZMODRHaTZXZzVaTHMxRDVWR21KWnUwN1pNaVJUcXVWVHhsTDBtOS16d0NwWnh0aDhPZnJFb1Y5SFIxc3BvMXJWTWQ5RldVdmhTN1dkR1ozRWFIOG5DeGJEMWpianVsWGNZaTljUjZ2MERYWk82ZW1UTjVkYTc4a1JZd2FjcFFsT0RYbmdfSVQ5ZnlpR0R2cWcwRnU4NmI5WHFNdkRqell3d3luOEFYdFlPZXBrMlJRUWdDa0s1VDljcThxY1p2dTAtQWdCVmozbkRKdXpmQXlodEltRmsyOUtPS2swd0V4Zkg0UVRDb0JreEhaVU8yRjZYVTR2cUExVWhkdGs2b1RzNnpCcWVB?oc=5" target="_blank">Exclusive | The Sudden Fall of OpenAI’s Most Hyped Product Since ChatGPT</a> <font color="#6f6f6f">WSJ</font>
Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models - WSJ
<a href="https://news.google.com/rss/articles/CBMiuANBVV95cUxNZUVvVzNvMVZBN09Hc3c5QWoxMUN4MEo1dDIzQjFXd1pkUlVWT25ZQ3pjOS1RMGdONzNrLUpfMHdLYTZUVm5YQkZXSDZxLU5qSkpCR0pYTTRURWtOR2JhMWMtMWVPX1dIRmlwN0lGTkFtVlVaRHdIeEFabFBTQU9FeWFiY3B1NERUNTc1X2N1djhGUk0yYVdUUjFzdGd0eFVuWVZ1TzN1akZCQmtuS3RzNWg3YVNraTFWd0ozX2hISTlUbElnQ29vUUx0WThlUVppU3E5LTNTSllDcXV0dUJUU29mWkVWZDV2OEtRbC1mRWYtWGZQOXUxZ1ppS1B6dThsSXM3V3lfNUxseERsUFVISW1HUXNUSVRhbUNLZ09sSjhDRnNGZXQtdGVsbGZCZXppRUdsdUhrWERta1V5Vm5GeWR6V1pvWE91TUFqQ0tfdU04TkpNTnIwSnd6RGNLWUo2ZVc1dHlzV1lWbi0xdWYtWXJCZ3RwTkdmS1R3Sk5RTl9iTFpTNXVFQktzYWZqYmNRZUpJMHBnRkNUb0tWMG1fV2JzckVhS1Zla0hoOE9Ra2hkdmZFT2lXWA?oc=5" target="_blank">Exclusive | Caltech Researchers Claim Radical Compression of High-Fidelity AI Models</a> <font color="#6f6f6f">WSJ</font>
Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!