You see the flour in the pantry, so you reach for it. You see the traffic light change to green, so you step on the gas. While the link between seeing and then moving in response is simple and essential to everyday existence, neuroscientists haven’t been able to get beyond debating where the link is and how it’s made. But in a new study in Nature Communications, a team at MIT’s Picower Institute for Learning and Memory provides evidence that one crucial brain region called the posterior parietal cortex (PPC) plays an important role in converting vision into action.
“Vision in the service of action begins with the eyes, but then that information has to be transformed into motor commands,” said senior author Mriganka Sur, Paul E. and Lilah Newton Professor of Neuroscience in the Department of Brain and Cognitive Sciences. “This is the place where that planning begins.”
Sur said the study may help to explain a particular problem in some people who have suffered brain injuries or stroke, called “hemispatial neglect.” In such cases, people are not able to act upon or even perceive objects on one side of their visual field. Their eyes and bodies are fine, but the brain just doesn’t produce the notion that there is something there to trigger action. Some studies have implicated damage to the PPC in such cases.
In the new study, the research team pinpointed the exact role of the PPC in mice and showed that it contains a mix of neurons attuned to visual processing, decision-making and action.
“This makes the PPC an ideal conduit for flexible mapping of sensory signals onto motor actions,” said Gerald Pho, a former graduate student in the Sur lab who is now at Harvard University. Pho is co-lead author with Michael Goard, a former MIT postdoctoral fellow now at UC Santa Barbara.
Where vision meets action. Credit: Sur Lab
Mouse see, mouse do
To do the research, the team trained mice on a simple task: if they saw a striped pattern on the screen drift upward, they could lick a nozzle for a liquid reward but if they saw the stripes moving to the side, they should not lick, lest they get a bitter liquid instead. In some cases they would be exposed to the same visual patterns, but the nozzle wouldn’t emerge. In this way, the researchers could compare the neurons’ responses to the visual pattern with or without the potential for motor action.
As the mice were viewing the visual patterns and making decisions whether to lick, the researchers were recording the activity of hundreds of neurons in each of two regions of their brain: the visual cortex, which processes sight, and the PPC, which receives input from the visual cortex, but also input from many other sensory and motor regions. The cells in each region were engineered to glow more brightly when they became active, giving the scientists a clear indication of exactly when they became engaged by the task.
Visual cortex neurons, as expected, principally lit up when the pattern appeared and moved, though they were split about evenly between responding to one visual pattern or the other.
Neurons in the PPC showed more varied responses. Some acted like the visual cortex neurons but most (about 70 percent) were active based on whether the pattern was moving the right way for licking (upward) and only if the nozzle was available. In other words, most PPC neurons were selectively responsive not merely to seeing something, but to the rules of the task and the opportunity to act on the correct visual cue.
“Many neurons in the PPC seemed to be active only during particular combinations of visual input and motor action,” Goard said. “This suggests that rather than playing a specified role in sensory or motor processing, they can flexibly link sensory and motor information to help the mouse respond to their environment appropriately.”
But even the occasional error was instructive. Consider the case when the nozzle was available and the stripes were moving sideways. In that case, a mouse should not lick even though it could. Visual cortex neurons behaved the same way regardless of the mouse’s decision, but PPC neurons were more active just before a mouse licked by mistake, than just before a mouse didn’t lick. This suggested that many PPC neurons are oriented toward acting.
Not yet fully convinced that the PPC encoded the decision to lick based on seeing the correct stripe movement, the researchers switched the rules of the task. Now, the nozzle would drip out the reward upon licking to the sideways stripe pattern and emit the bitter liquid when the stripes moved up. In other words, the mouse still saw the same things, but what they meant for action had reversed.
With the same mice re-trained, the researchers then looked again at the same neurons in the same regions. Visual cortex neurons didn’t change their activity at all. Those that followed the upward pattern or sideways pattern still did as before. What the mice were seeing, after all, hadn’t changed.
However, the neural responses in PPC changed along with the rules for action. Neurons that had been activated selectively for upward visual patterns now responded instead to sideways patterns. In other words, learning of the new rules was directly evident in the changed activity of neurons in the PPC. The researchers therefore observed a direct correlate of learning at the cellular level, strongly implicating the PPC as a critical node for where seeing meets acting on that information.
“If you flipped the rules of traffic lights so that red means go, the visual input would still be driven by the colors, but the linkage to motor output neurons would switch, and that happens in the PPC,” Sur said.
The findings extend earlier results made by other researchers in primates, the researchers wrote, suggesting that mice bear the needed similarity to aid further studies of the PPC.
“Our understanding of how decisions are computed and visuomotor transformations are made, will be greatly aided by future circuit-level analyses of PPC function in this powerful model system,” they concluded.
This article has been republished from materials provided by Picower Institute, MIT. Note: material may have been edited for length and content. For further information, please contact the cited source.
Pho, G., Goard, M., Woodson, J., Crawford, B., & Sur, M. (2018). Task-dependent representations of stimulus and choice in mouse parietal cortex. Nature Communications, 9(1). doi: 10.1038/s41467-018-05012-y