A press release
from UCLA highlights the work of some of their researchers, which is published in the 10 September 2004 issue
Left and Right Ears Not Created Equal as Newborns Process Sound, UCLA/University of Arizona Scientists Discover
Date: September 9, 2004
Contact: Elaine Schmidt ( email@example.com )
Challenging decades of scientific belief that the decoding of sound originates from a preferred side of the brain, UCLA and University of Arizona scientists have demonstrated that right-left differences for the auditory processing of sound start at the ear.
Reported in the Sept. 10 edition of Science, the new research could hold profound implications for rehabilitation of persons with hearing loss in one or both ears, and help doctors enhance speech and language development in hearing-impaired newborns.
"From birth, the ear is structured to distinguish between various types of sounds and to send them to the optimal side in the brain for processing," said Yvonne Sininger, visiting professor of head and neck surgery at the David Geffen School of Medicine at UCLA. "Yet no one has looked closely at the role played by the ear in processing auditory signals."
Scientists have long understood that the auditory regions of the two halves of the brain sort out sound differently. The left side dominates in deciphering speech and other rapidly changing signals, while the right side leads in processing tones and music. Because of how the brain's neural network is organized, the left half of the brain controls the right side of the body, and the left ear is more directly connected to the right side of the brain.
Prior research had assumed that a mechanism arising from cellular properties unique to each brain hemisphere explained why the two sides of the brain process sound differently. But Sininger's findings suggest that the difference is inherent in the ear itself.
"We always assumed that our left and right ears worked exactly the same way," she said. "As a result, we tended to think it didn't matter which ear was impaired in a person. Now we see that it may have profound implications for the individual's speech and language development."
Working with co-author Barbara Cone-Wesson, associate professor of speech and hearing sciences at the University of Arizona, Sininger studied tiny amplifiers in the outer hair cells of the inner ear.
"When we hear a sound, tiny cells in our ear expand and contract to amplify the vibrations," Sininger said. "The inner hair cells convert the vibrations to neural cells and send them to the brain, which decodes the input."
"These amplified vibrations also leak back out to the ear in a phenomena call otoacoustic emission (OAE)," Sininger said. "We measured the OAE by inserting a microphone in the ear canal."
In a six-year study, the UCLA/University of Arizona team evaluated more than 3,000 newborns for hearing ability before they left the hospital. Sininger and Cone-Wesson placed a tiny probe device in the baby's ear to test its hearing. The probe emitted a sound and measured the ear's OAE.
The researchers measured the babies' OAE with two types of sound. First, they used rapid clicks and then sustained tones. They were surprised to find that the left ear provides extra amplification for tones like music, while the right ear provides extra amplification for rapid sounds timed like speech.
"We were intrigued to discover that the clicks triggered more amplification in the baby's right ear, while the tones induced more amplification in the baby's left ear," Sininger said. "This parallels how the brain processes speech and music, except the sides are reversed due to the brain's cross connections."
"Our findings demonstrate that auditory processing starts in the ear before it is ever seen in the brain," Cone-Wesson said. "Even at birth, the ear is structured to distinguish between different types of sound and to send it to the right place in the brain."
[ ... Read the full report ...]