Foundation Models for Neuroscience

Apr 19, 2025

Neural foundation models are coming (some are already here). But are we ready for what they actually require—and what they really deliver?

Foundation what?

Unless you’ve been living under a rock, you’ve probably heard of foundation models. They’re large AI systems trained on massive datasets that generalize to a wide range of tasks. You train them once, and then fine-tune or prompt them to do just about anything. ChatGPT is one of the most familiar examples. It’s a chatbot built on top of a foundation model (GPT-3, GPT-4, etc.), trained to predict the next token in a sequence of text. This approach embodies the “Bitter Lesson” where a remarkably simple objective (predict the next token) combined with an enormous amount of data (and a really impressive engineering stack to process it), leads to something, well, incredible.

This approach has delivered some huge wins in AI and science. The Nobel Prize in chemistry in 2024 went to three scientists who used AlphaFold, a foundation model for protein structure, to predict the structure of dozens of proteins that were previously unsolved. I’d venture that a huge swath of scientists are spending time figuring out how to apply this approach to their own domains. The key idea is simple: format your scientific question problem as as a loss function you can optimize, throw a ton of data and compute at it, and you’ll get something useful out.

For many neuroscientists, this is already how we think. We’ve been fitting data-driven descriptive models to neurons using numerical optimization for decades. Unsurprisingly, there are broad calls to follow the lead of AlphaFold and scale up this approach in neuroscience.

The responses to the Dyer and Richards article are worth reading, especially the one from David Sussillo, but can we build a foundation model for neuroscience? and, what would it give us?

The answer to the first question is “probably”, but maybe we should be cautious.

Problem 1: Data

Foundation models are only as good as the data they’re trained on. And neuroscience data is… not great.

Recording from brains is hard. Every tool has major limitations. fMRI? It’s a slow measure of blood oxygenation, not neural computation. Two-photon imaging? Beautiful spatial detail, but slow and artifact-prone. Electrophysiology? The gold standard, but sample sizes are small, subsampled, and increasingly poor quality.

Even our best open datasets—like from IBL or the Allen Institute—are highly specific : one species, one setup, one behavior. There’s no shared protocol, no aligned vocabulary. Foundation models thrive on scale and alignment. Neuroscience has neither.

What foundation models actually give you

Here’s the other thing that often gets missed in the hype: ChatGPT didn’t solve language. It didn’t explain syntax, or how kids learn grammar, or why words mean what they do. But it gave us what solving language would have given us: the ability to generate, translate, summarize, and converse fluently.

It gave us operational understanding—the ability to do useful things without necessarily knowing why they work.

That’s powerful. And if we can pull it off in neuroscience, it would be transformative. Imagine models that can decode intent, predict mental state, guide a neuroprosthetic, detect disorders before symptoms appear. These don’t require scientific understanding—they just require prediction that works.

That’s enough to change lives. And it’s why foundation models are still worth chasing.

But prediction ≠ understanding

Still, science isn’t just about getting the answer—it’s about understanding the system.

A model that predicts spikes well isn’t necessarily telling you how the brain computes. It might just be picking up on correlations in the data. It could be exploiting quirks of the task, the subject, or the stimulus. And unless we open the box and understand the computation inside, we won’t be any closer to a theory of the brain.

That’s the risk: we end up building black boxes that work, but teach us nothing.

There’s an alternative. Use these large models as experimental systems. Train them to perform the same tasks as animals. Analyze their internal structure. Compare their representations and dynamics to real circuits. Use them to generate hypotheses. In short: reverse-engineer them. That’s where the insight lives.

Learn from the limits of LLMs

If we don’t proceed carefully, we’ll repeat the mistakes of large language models.

LLMs hallucinate. They confidently generate false facts. They overfit quirks in training data. They reflect bias, amplify noise, and often generalize poorly outside their domain.

In neuroscience, the analogs are easy to imagine. A model that predicts spikes accurately, but only because it’s learned the eye movement pattern of a particular animal. A decoder that only works on one task, in one lab. A BCI that breaks down when arousal state changes.

That’s not robustness. That’s fragility at scale.

Where we go from here

So what’s the path forward?

It’s not to abandon foundation models. It’s to earn them. That means building the conditions that made GPT and AlphaFold possible: big, aligned datasets. Open infrastructure. Shared benchmarks. Teams of engineers and scientists working together.

It also means staying honest about what we’re doing. Operational tools can be transformative. But if we want science—not just applications—we need models we can interrogate, not just deploy.

Foundation models could help us treat disease, decode thought, and restore movement. But if we want to understand the brain—truly understand it—we’ll need more than predictions. We’ll need interpretation, theory, and better data.

It’s not one or the other. It’s both.