We've updated our Privacy Policy to make it clearer how we use your personal data. We use cookies to provide you with a better experience. You can read our Cookie Policy here.

Advertisement

Brain appears to have different mechanisms for reconciling sight and sound

Listen with
Speechify
0:00
Register for free to listen to this article
Thank you. Listen to this article using the player above.

Want to listen to this article for FREE?

Complete the form below to unlock access to ALL audio articles.

Read time: 3 minutes

Study sheds light on how we process visual and aural information -


A new University of California, Los Angeles (UCLA) psychology study provides insights into how the brain combines sound and vision. The research suggests that there is not one sole mechanism in the brain that governs how much our senses work together to process information.


Among the implications of the study: It might not be as easy as many people had assumed to categorize the way in which we perceive and learn.


See Also: Study helps fill in gaps in our visual perception


"We should be cautious not to make blanket statements about how we process information, like 'I'm a visual learner,'" said Ladan Shams, an associate professor of psychology in the UCLA College and senior author of the research. "That's not necessarily true across the board. For example, your brain may combine sights and sounds a lot in one task—watching TV, for example—but only a little in another task—such as playing the piano."


The researchers found that people's vision frequently influenced their hearing when they tried to identify the specific location of sounds and flashes of light, and that their hearing influenced vision when they counted the sounds and flashes.


In one part of the study, 59 participants, mostly UCLA undergraduates, were seated in front of a computer monitor with speakers on either side and asked to count the number of flashes of light on the screen and beeps played on the speakers. Sometimes they only saw flashes, sometimes they only heard beeps and sometimes they both saw flashes and heard beeps—in which case the numbers could vary, up to four of each. The researchers presented 360 combinations of beeps and flashes in a one-hour period.


"When people have to process different numbers of beeps and flashes, it's really hard—the senses blend together," said Brian Odegaard, a UCLA postdoctoral scholar who was the study's lead author. "Most people, when presented with two beeps and one flash, mistakenly said they saw two flashes, while a few participants could accurately tease apart lights and sounds."


Shams explained that most people are pretty good at identifying visual stimuli if no other senses are involved, but in the experiment, what participants saw was typically affected by what they heard.


"When there's sound, many people get fooled into thinking that there are fewer or more flashes," she said.


Learn More: Perception: Brain integrates features directly to patterns


In the second part of the study, subjects were asked to sit facing a black screen, behind which were five speakers. A projector mounted overhead was used to flash bursts of light onto the screen, at the same spots where the speakers were located.


The researchers played brief bursts of sound and triggered flashes of light, in various combinations, and asked participants to identify where they originated. The participants used a computer mouse to point at the part of the screen where they thought the sounds and flashes occurred; each was presented with 525 trials.


When the flash and sound occurred in different locations from each other, most participants struggled to correctly identify where the sound occurred. The effect was similar to what happens when people watch a ventriloquist with a puppet or a magician using misdirection to pull off an illusion, Odegaard said.


Given that the two tasks both involved brief sights and sounds, the researchers figured that the people who were fooled less on one task would be fooled less on the other as well, but the researchers were surprised to find no correlation in how much subjects combined the two senses on the two tasks.


A week later, the participants were given both of the same tasks again. The results were the same—although they performed consistently on each individual task, their abilities to bind the two senses in the two different tests showed no correlation.3


Related: Taste perception is influenced by extreme noise conditions


People vary in how much their hearing and vision interact, Shams explained. For example, some find it easier to understand someone who is speaking when they are looking at the speaker. Some are better able, when driving, to estimate how fast other vehicles are traveling using both visual and aural cues.


"This finding suggests that if you find a strategy of combining your senses that helps you perform a given task, stick with it," she said.


This study was the first to test the same group of people in both a spatial task and counting task. In future research, Shams and Odegaard hope to discover whether we can learn to modify how our brain combines sounds and sights, and if so, how.


Note: Material may have been edited for length and content. For further information, please contact the cited source.

University of California, Los Angeles   Original reporting by: Stuart Wolpert


Publication

Odegaard B, Shams L. The Brain’s Tendency to Bind Audiovisual Signals Is Stable but Not General. Psychological Science, Published Online March 4 2016. doi: 10.1177/0956797616628860