Concentrating on word sounds helps reading instruction and intervention
News Jan 26, 2015
Findings point to the value of word sounds over visual processing during reading instruction or when diagnosing and treating reading disorders
A neuroimaging study by a University at Buffalo psychologist suggests that phonics, a method of learning to read using knowledge of word sounds, shouldn't be overlooked in favor of a whole-language technique that focuses on visually memorizing word patterns, a finding that could help improve treatment and diagnosis of common reading disorders such as dyslexia.
"Phonological information is critical for helping identify words as they're being read," says Chris McNorgan, PhD, assistant professor of psychology, whose study, "Skill dependent audiovisual integration in the fusiform induces repetition suppression," used MRI scans to observe how parts of the brain responded to audio and visual word cues. The results are published in the most recent edition of Brain & Language.
A better reader is someone whose visual processing is more sensitive to audio information, according to the study's results.
"There are applications here not just for reading disorders, but also for how children are taught to read in the classroom," he says.
Barring injury, McNorgan says, all parts of the brain are working at all times, contrary to the myth that it functions at only a fraction of its capacity. However, different parts of the brain are specialized for different types of activities that trigger some regions to work harder than others.
With reading, the Visual Word Form Area (VWFA) is excited when it encounters familiar letter combinations. But most activities require communication between different brain regions and coordination with sensory systems, like an outfielder watching a baseball while the brain programs the motor system to catch it.
How this communication happens while reading -- which requires visual and auditory knowledge -- and to what extent is less clear. So McNorgan's study looked for what's known as top-down influence of auditory knowledge in the VWFA.
Think of a bottom-up process as a flow of information that begins with the visual system feeding neurons that detect basic features in words such as line orientation that eventually leads to word recognition. A top-down process implies that some other information enters that flow of visual recognition -- information like the knowledge of the word sounds.
"This auditory knowledge can be used to help rule out some letter combinations. For example, many words end in ISK or ASK. For a few milliseconds there may be some ambiguity among the neurons trying to figure out whether that last letter is a K or an X," said McNorgan. "Since you don't have any words ending in ISX in your verbal repertoire, this helps rule out the possibility that you read the word DISX and instead read the word as DISK."
To find evidence of this top-down input, researchers presented subjects with wide ranges of reading abilities between the ages of 8 and 13 with word pairs. The subjects had to determine if the words rhymed while an MRI scanner monitored their brain activity.
The experiment used three sets of conditions when presenting the word pairs: subjects first read the word pairs (visual-only); then heard the word pairs (auditory-only); and lastly, a combination of sight and sound, hearing the first word but reading the second (audio-visual). The MRI scanner determined which parts of the brain were most active during each condition by displaying a three dimensional representation of the brain, made up of what look like a series of cubes, called voxels.
"Think of the voxels as LEGOS assembled together to make a 3D model of the brain. Each cube has a measurement of activation strength that allows us to understand of what's happening in each area under all three of the conditions," said McNorgan.
The resulting images, he said, comprise something like a movie reel, with approximately one frame passing every two seconds. Signal strength is then measured in each voxel under all the condition across all the snapshots in time.
"Looking at the voxels in a particular brain area, if the signal strengths associated with two different conditions differ, then you have some evidence that brain area processes information about the two conditions differently," says McNorgan.
To make sense of the results through all the conditions, researchers take the sum of the auditory-only and visual-only signals and compare that to the strength of the audio-visual condition. This helps them distinguish between multisensory sensory neurons, which become excited by audio-visual information, and collections of heterogeneous unisensory neurons, a mix of visual-only and auditory-only that respond excitedly to one or the other.
"If the audio-visual response is greater than the sum of the auditory-only and the visual-only, this suggests that getting both types of inputs causes these neurons to fire for longer periods of time. This is a superadditive effect," says McNorgan. "An audio-visual response less than that sum suggests that getting both types of inputs causes these neurons to fire for less time. This is a subadditive effect."
This subadditivity is associated with higher reading scores and faster responses to similarly spelled words, the reading equivalent to having a head start in a race. "As you learn how to read, your brain starts to make more use of top-down information about the sounds of letter combinations in order to recognize them as parts of words," says McNorgan. "This information gives your word-recognition system a leg-up, allowing it to respond more quickly. The multisensory neurons are getting the job done sooner, so they don't need to fire for as long. Better readers seem to have more of these neurons taking advantage of auditory information to help the visual word recognition system along."
Early intervention and basic instruction would counterintuitively involve this auditory information, "thinking more about the sounds of different words instead of concentrating on recognizing words," says McNorgan.
Note: Material may have been edited for length and content. For further information, please contact the cited source.
Chris McNorgan, James R. Booth. Skill dependent audiovisual integration in the fusiform induces repetition suppression. Brain and Language, Published February 2015. doi: 10.1016/j.bandl.2014.12.002
Some of the most advanced tools in neuroscience are developed for use with mice, yet studies of behaviors most relevant to humans typically involve other model organisms. Anne Churchland and colleagues addressed this gap by investigating the mouse's potential as an animal model of decision-making.READ MORE
When MIT neuroscientist Earl Miller was in graduate school at Princeton, he was inspired by the lectures of George A. Miller, an influential psychologist who helped to spark the young student’s interest in working memory. Now, as the newly named 2019 recipient of the George A. Miller Prize in Cognitive Neuroscience, Earl Miller is set to deliver a lecture honoring his teacher.READ MORE