How the Brain Fuses the Senses: A Classic Paper That Changed How We Think About Perception
That ability to combine information from different senses is so seamless that we rarely notice it. But understanding how the brain does this turns out to be one of the most important questions in neuroscience.
In 2008, neuroscientists Barry Stein and Terrence Stanford published what is now considered a foundational paper on multisensory integration — the process by which the brain combines signals from sight, sound, touch, and other senses into unified, meaningful experiences. Rather than focusing on perception in the abstract, this paper zoomed in on something surprisingly concrete: what individual neurons actually do when they receive information from more than one sense.
What they uncovered reshaped how scientists think about perception, attention, development, and even disorders of sensory processing.
Let’s unpack the ideas — using the real scientific terms, but in plain language.
What is multisensory integration, really?
Multisensory integration (MSI) refers to the brain’s ability to combine information from different sensory modalities — like vision and hearing — in a way that improves detection, speed, or accuracy.
Crucially, this is not just about having multiple senses active at the same time.
From a neuroscience perspective, multisensory integration means this:
A neuron responds differently to a combined stimulus (for example, sight + sound) than it does to the strongest single stimulus alone.
That difference is what makes MSI interesting. It shows the brain is doing nonlinear computation, not just adding things up.
Meet the multisensory neuron
A multisensory neuron is a neuron that responds to — or is influenced by — more than one sensory modality.
For example:
-
A neuron might fire when it sees something,
-
fire when it hears something,
-
and fire even more strongly when it both sees and hears the event.
This “extra” response is not automatic. It follows rules.
Enhancement, depression, and why “more” isn’t always more
When multiple senses are combined, a neuron’s response can change in two main ways:
-
Multisensory enhancement:
The neuron responds more strongly to the combined stimulus than to the strongest single-sense stimulus. -
Multisensory depression:
The combined response is actually weaker than the strongest single-sense response.
Both are forms of multisensory integration.
Why would the brain ever suppress information when more senses are involved? Because integration isn’t about piling on signals — it’s about deciding what matters.
Superadditive, additive, and subadditive: three ways neurons combine senses
One of the most important contributions of this paper is the idea that multisensory integration comes in three computational flavors:
-
Superadditive
The combined response is greater than the sum of the individual responses.
This is the biggest “boost” and happens when signals are weak. -
Additive
The combined response is roughly equal to the sum. -
Subadditive
The combined response is less than the sum — sometimes barely more than one signal alone.
This leads to a key principle called inverse effectiveness.
Inverse effectiveness: why weak signals benefit the most
Inverse effectiveness means this:
The weaker the individual sensory signals, the bigger the relative benefit of combining them.
If a sound is loud and a visual signal is clear, your brain doesn’t gain much by merging them — each is already informative. But if both signals are faint or ambiguous, combining them can dramatically improve detection.
This principle explains why multisensory integration is especially important in:
-
low visibility,
-
noisy environments,
-
early development,
-
and many clinical conditions.
Why location matters: the spatial rule
The brain assumes that signals belong together only if they come from the same place.
In the paper, Stein and Stanford show that multisensory neurons follow a strict spatial principle:
-
Visual and auditory receptive fields (the regions of space a neuron responds to) must overlap
-
If a sound comes from one location and a visual signal from another, integration weakens or reverses
This makes sense from an evolutionary perspective: if cues don’t align in space, they might represent different events.
Timing matters too: the temporal rule
Even signals from the same place won’t be integrated if they are too far apart in time.
The brain uses a temporal binding window — a period during which signals can be linked despite different sensory delays (for example, sound travels slower than light).
Integration is strongest when neural responses overlap in time, not just when stimuli occur simultaneously.
The superior colliculus: a multisensory hub for action
Much of the paper focuses on a midbrain structure called the superior colliculus (SC).
The SC is involved in:
-
orienting the eyes and head,
-
shifting attention,
-
rapidly responding to important events.
It turns out the SC is packed with multisensory neurons — making it a perfect place to study how integration works at the single-neuron level.
Importantly, multisensory integration in the SC directly improves behavior: faster responses, better localization, quicker reactions.
Cortex is essential — even for midbrain integration
One of the paper’s most striking findings is that multisensory integration in the superior colliculus depends on the cortex.
When researchers temporarily deactivate certain association cortical areas:
-
SC neurons still respond to multiple senses,
-
but they lose multisensory enhancement,
-
and the behavioral benefits disappear.
This shows that multisensory integration is not a simple reflex — it’s a distributed circuit process involving dialogue between cortex and midbrain.
Multisensory integration is learned, not innate
Perhaps most surprisingly, the ability to integrate senses is not present at birth.
In animal studies:
-
newborn neurons respond to multiple senses,
-
but they cannot integrate them effectively,
-
integration develops only with experience.
When animals are raised without normal cross-sensory experience (for example, in darkness), multisensory integration fails to develop — or develops incorrectly.
This makes multisensory integration a powerful example of experience-dependent brain plasticity.
Cortex does things differently
In higher cortical areas, multisensory integration becomes more complex.
Instead of focusing only on where and when, cortical regions care about:
-
meaning,
-
context,
-
semantic congruence.
For example, brain areas involved in communication respond more strongly when facial movements match vocal sounds — and may suppress responses when they don’t.
This shows that multisensory integration is not a single process, but a family of processes shaped by the goals of each brain region.
Is there really such a thing as “unisensory” cortex?
The paper ends by raising a provocative question: if even early sensory areas receive input from other senses, should we still call them “visual cortex” or “auditory cortex”?
The authors take a cautious stance:
-
Traditional sensory labels still matter,
-
but multisensory influences are more widespread than once thought.
Rather than abolishing the idea of unisensory cortex, they suggest recognizing transitional and integrative zones.
Why this paper still matters
Even 2 decades later, this paper remains foundational because it showed that multisensory integration:
-
is nonlinear,
-
follows clear rules,
-
depends on development and experience,
-
emerges from distributed neural circuits,
-
and directly shapes perception and behavior.
It laid the groundwork for modern research on:
-
attention,
-
peripersonal space,
-
predictive processing,
-
sensory differences in autism and ADHD,
-
and how the brain constructs a coherent world from noisy inputs.
In short, it taught us that perception is not about passive reception — it’s about active synthesis.
Stein, B. E., & Stanford, T. R. (2008). Multisensory integration: current issues from the perspective of the single neuron. Nature reviews. Neuroscience, 9(4), 255–266. https://doi.org/10.1038/nrn2331

