Neurons across eight areas of the brain team up to process visual information. Researchers have now documented these neurons’ activity at a breadth and resolution never before reported.
Using a custom-built microscope to peer into the mouse brain, scientists have tracked the activity of single neurons across the entire visual cortex.
These recordings, made in the tenths of seconds after the animals saw a cue on a screen, expose the complex dynamics involved in making sense of what the eyes see. In an unprecedented combination of breadth and detail, the results describe the behavior of more than 21,000 total neurons in six mice over five days, Howard Hughes Medical Institute Investigator Mark Schnitzer’s team reports in the journal Nature on May 18, 2022.
His team is the first to get a glimpse of individual cells’ activity occurring at the same time throughout eight parts of the brain involved in vision. “People have studied these brain areas before, but prior imaging studies did not have cellular resolution across the entire visual cortex,” says Schnitzer, a neuroscientist at Stanford University.
The work highlights the dramatic sequence of events that unfolds in the brain from the instant it receives messages from the eyes until it decides how to respond to that sight. The researchers’ far-reaching but fine-grained imaging approach made it possible for them to collect an “incredible” set of data, says Tatiana Engel, a computational neuroscientist at Cold Spring Harbor Laboratory who was not involved in the study.
Mark Schnitzer, HHMI Investigator at Stanford University
While previous studies have already explored aspects of this process, such as variations in single neurons’ activity and coordination between larger brain areas, this research offers an expansive new view, she says. “The scale on which they’re able to address these topics is very impressive.”
When the eyes see an image, they send electrical signals that end up in the visual cortex, the wrinkly outer layer of the brain near the back of the head. There, the signals trigger a flurry of activity as neurons work together to register an image, evaluate it, and decide how to respond.
To capture activity across the visual cortex, Schnitzer and his colleagues built a custom microscope with a wide field of view. Their system could also capture detail at a resolution of a few thousandths of a millimeter, small enough to detect single neurons. By using genetically engineered mice with neurons that fluoresce when sending signals, the team could watch these cells’ activity.
During the team’s experiments, mice had to make a choice based on one of two visual cues. One prompted the animals to lick a spout for some sugar water, the other cue indicated “don’t lick.” The mice performed many of these tests over five days.
With recordings made from the mice’s brains, the team posed a simple question: What happens in the brain when we see something? Their results lay out this invisible process at a time-resolution of fractions of a second and uncover surprising nuances.
Scientists, for example, already knew that individual neurons behave variably when responding to visual signals conveyed by the eyes. But Schnitzer’s team’s experiments revealed a pattern to this unreliable behavior. That pattern could make it easier for brain areas receiving the neurons’ signals to make sense of them and accurately interpret the visual scene.
The researchers also documented how, about 200 milliseconds after the visual cue appeared, the animals switched mental gears: messages from the eyes prompted a massive rearrangement in different brain areas’ activity. About 500 milliseconds afterward, this surge subsided and the activity became more stable and recognizable. Next, roughly 600 milliseconds later, another signal emerged, activating all eight of the brain areas. That signal encoded the animal’s decision to stay still or go for the sugar water. The researchers learned how to read the signal, so they could predict which response the mouse would make.
“It’s fascinating how much the brain is doing in the immediate moments after the eyes see the stimulus,” Schnitzer says.
Sadegh Ebrahimi et al. “Emergent reliability in sensory cortical coding and inter-area communication.” Nature. Published online May 18, 2022. doi: 10.1038/s41586-022-04724-y