4.5 Audition: More Than Meets the Ear

Vision is based on the spatial pattern of light hitting the retina. The sense of hearing, by contrast, is all about sound waves: changes in air pressure unfolding over time. Plenty of things produce sound waves: the collision of a tree hitting the forest floor, the impact of two hands clapping, the vibration of vocal cords during a stirring speech, the resonance of a bass guitar string during a thrash metal concert. Understanding auditory experience requires understanding how we transform changes in air pressure into perceived sounds.

4.5.1 Sensing Sound

Striking a tuning fork produces a pure tone, a simple sound wave that first increases air pressure and then creates a relative vacuum. This cycle repeats hundreds or thousands of times per second as sound waves travel outward in all directions from the source. Just as there are three dimensions of light waves corresponding to three dimensions of visual perception, so, too, there are three physical dimensions of a sound wave. Frequency, amplitude, and complexity determine what we hear as the pitch, loudness, and quality of a sound (see TABLE 4.4).

Most sounds—such as voices, music, trees rustling in the wind, the screech of brakes, the purring of a cat—are composed of not one, but many different frequency components. Although you perceive the mixture (the cat’s purr, for example, not the 206 Hz component in the purr), the first thing the ear does to a sound is to analyze it—break it down into its separate component frequencies. The psychological attributes of pitch, loudness, and timbre are then “built up” by the brain from the separate frequency components that are represented in the inner ear, just like visual perceptions are “built up” from the spatial pattern of activity on the retina. The focus in our discussion of hearing, then, is on how the auditory system encodes and represents sound frequency (Kubovy, 1981).

Table 4.4: Properties of Sound Waves

159

4.5.2 The Human Ear

LEO CULLUM/THE NEW YORKER COLLECTION/ CARTOONBANK.COM

How does the auditory system convert sound waves into neural signals? The process is very different from the visual system, which is not surprising given that light is a form of electromagnetic radiation, whereas sound is a physical change in air pressure over time: Different forms of energy require different processes of transduction. The human ear is divided into three distinct parts, as shown in FIGURE 4.24. The outer ear collects sound waves and funnels them toward the middle ear, which transmits the vibrations to the inner ear, embedded in the skull, where they are transduced into neural impulses.

The outer ear consists of the visible part on the outside of the head (called the pinna); the auditory canal; and the eardrum, an airtight flap of skin that vibrates in response to sound waves gathered by the pinna and channelled into the canal. The middle ear, a tiny, air-filled chamber behind the eardrum, contains the three smallest bones in the body, called ossicles. Named for their appearance as hammer, anvil, and stirrup, the ossicles fit together into a lever that mechanically transmits and intensifies vibrations from the eardrum to the inner ear.

How do hair cells in the ear enable us to hear?

The inner ear contains the spiral-shaped cochlea (Latin for “snail”), a fluid-filled tube that is the organ of auditory transduction. The cochlea is divided along its length by the basilar membrane a structure in the inner ear that undulates when vibrations from the ossicles reach the cochlear fluid (see FIGURE 4.25). Its wavelike movement stimulates thousands of tiny hair cells, specialized auditory receptor neurons embedded in the basilar membrane. The hair cells then release neurotransmitter molecules, initiating a neural signal in the auditory nerve that travels to the brain. You might not want to think that the whispered “I love you” that sends chills up your spine got a kick start from lots of little hair cells wiggling around, but the mechanics of hearing are what they are!

Figure 4.24: Anatomy of the Human Ear The pinna funnels sound waves into the auditory canal to vibrate the eardrum at a rate that corresponds to the sound’s frequency. In the middle ear, the ossicles pick up the eardrum vibrations, amplify them, and pass them along by vibrating a membrane at the surface of the fluid-filled cochlea in the inner ear. Here fluid carries the wave energy to the auditory receptors that transduce it into electrochemical activity, exciting the neurons that form the auditory nerve, leading to the brain.
Figure 4.25: Auditory Transduction Inside the cochlea (shown here as though it were uncoiling) the basilar membrane undulates in response to wave energy in the cochlear fluid. Waves of differing frequencies ripple varying locations along the membrane, from low frequencies at its tip to high frequencies at the base, and bend the embedded hair cell receptors at those locations. The hair cell motion generates impulses in the auditory neurons, whose axons form the auditory nerve that emerges from the cochlea.

160

4.5.3 Perceiving Pitch

Figure 4.26: Primary Auditory Cortex Area A1 is folded into the temporal lobe beneath the lateral fissure in each hemisphere. The left hemisphere auditory areas govern speech in most people. A1 cortex has a topographic organization (inset), with lower frequencies mapping toward the front of the brain and higher frequencies toward the back, mirroring the organization of the basilar membrane along the cochlea (see Figure 4.25).

From the inner ear, action potentials in the auditory nerve travel to the thalamus and ultimately to an area of the cerebral cortex called area A1, a portion of the temporal lobe that contains the primary auditory cortex (see FIGURE 4.26). For most of us, the auditory areas in the left hemisphere analyze sounds related to language and those in the right hemisphere specialize in rhythmic sounds and music. There is also evidence that the auditory cortex is composed of two distinct streams, roughly analogous to the dorsal and ventral streams of the visual system. Spatial (“where”) auditory features, which allow you to locate the source of a sound in space, are handled by areas toward the back (caudal) part of the auditory cortex, whereas nonspatial (“what”) features, which allow you to identify the sound, are handled by areas in the lower (ventral) part of the auditory cortex (Recanzone & Sutter, 2008). Note: All of hearing, including perceiving “where,” depends on “temporal aspects.”

161

Neurons in area A1 respond well to simple tones, and successive auditory areas in the brain process sounds of increasing complexity (see Figure 4.26, inset) (Rauschecker & Scott, 2009; Schreiner, Read, & Sutter, 2000; Schreiner & Winer, 2007). The human ear is most sensitive to frequencies around 1000 to 3500 Hz. But how is the frequency of a sound wave encoded in a neural signal? Our ears have evolved two mechanisms to encode sound-wave frequency, one for high frequencies and one for low frequencies.

4.5.4 Localizing Sound Sources

Just as the differing positions of our eyes give us stereoscopic vision, the placement of our ears on opposite sides of the head gives us stereophonic hearing. The sound arriving at the ear closer to the sound source is louder than the sound in the farther ear, mainly because the listener’s head partially blocks sound energy. This loudness difference decreases as the sound source moves from a position directly to one side (maximal difference) to straight ahead (no difference).

162

Another cue to a sound’s location arises from timing: Sound waves arrive a little sooner at the near ear than at the far ear. The timing difference can be as brief as a few microseconds but, together with the intensity difference, it is sufficient to allow us to perceive the location of a sound. When the sound source is ambiguous, you may find yourself turning your head from side to side to localize it. By doing this, you are changing the relative intensity and timing of sound waves arriving in your ears and collecting better information about the likely source of the sound. Turning your head also allows you to use your eyes to locate the source of the sound—and your visual system is much better at pinpointing the location of things than your auditory system is.

4.5.5 Hearing Loss

Sound amplification helps in the case of which type of hearing loss?

Broadly speaking, hearing loss has two main causes. Conductive hearing loss arises because the eardrum or ossicles are damaged to the point that they cannot conduct sound waves effectively to the cochlea. The cochlea itself, however, is normal, making this a kind of “mechanical problem” with the moving parts of the ear: the hammer, anvil, stirrup, or eardrum. In many cases, medication or surgery can correct the problem. Sound amplification from a hearing aid also can improve hearing through conduction via the bones around the ear directly to the cochlea.

A cochlear implant works when a microphone picks up sounds and sends them to a small speech-processing computer worn on the user’s belt or behind the ear. The electric signals from the speech processor are transmitted to an implanted receiver, which sends the signals via electrodes to the cochlea, where the signals directly stimulate the auditory nerve.
AP PHOTO/ROCHESTER POST-BULLETIN, JERRY OLSON

Sensorineural hearing loss is caused by damage to the cochlea, the hair cells, or the auditory nerve, and it happens to almost all of us as we age. Sensorineural hearing loss can be heightened in people regularly exposed to high noise levels (such as rock musicians or jet mechanics). Simply amplifying the sound does not help because the hair cells can no longer transduce sound waves. In these cases a cochlear implant may offer some relief.

A cochlear implant is an electronic device that replaces the function of the hair cells (Waltzman, 2006). The external parts of the device include a microphone and speech processor, about the size of a USB key, worn behind the ear, and a small, flat, external transmitter that sits on the scalp. The implanted parts include a receiver just inside the skull and a thin wire containing electrodes inserted into the cochlea to stimulate the auditory nerve. Sound picked up by the microphone is transformed into electric signals by the speech processor, which is essentially a small computer. The signal is transmitted to the implanted receiver, which activates the electrodes in the cochlea. Cochlear implants are now in routine use and can improve hearing to the point where speech can be understood.

Marked hearing loss is commonly experienced by people as they grow older, but is rare in an infant. However, infants who have not yet learned to speak are especially vulnerable because they may miss the critical period for language learning (see the Learning chapter). Without auditory feedback during this time, normal speech is nearly impossible to achieve, but early use of cochlear implants has been associated with improved speech and language skills for hearing-impaired children (Hay-McCutcheon et al., 2008). Efforts are under way to introduce cochlear implants to children as young as 12 months old or younger to maximize their chances of normal language development (DesJardin, Eisenberg, & Hodapp, 2006; Holman et al., 2013). (For research on the importance of music and brain development, see The Real World box.)

163

THE REAL WORLD: Music Training: Worth the Time

Did you learn to play an instrument when you were younger? Maybe for you, music was its own reward (or maybe not). For anyone who practised mostly to keep Mom and Dad happy, there is good news. Musical training has a range of benefits. Let us start with the brain. Musicians have greater plasticity in the motor cortex compared with nonmusicians (Rosenkranz et al., 2007); they have increased grey matter in motor and auditory brain regions compared with nonmusicians (Gaser & Schlaug, 2003; Hannon & Trainor, 2007); and they show differences in brain responses to musical stimuli compared with nonmusicians (Pantev et al., 1998). But musical training also extends to auditory processing in nonmusical domains (Kraus & Chandrasekaran, 2010). Musicians, for example, show enhanced brain responses when listening to speech compared with nonmusicians (Parbery-Clark et al., 2012). Musicians also exhibit an improved ability to detect speech when it is presented in a noisy background (Parbery-Clark, Skoe, & Kraus, 2009). This effect has been demonstrated in children, young adults, and even older adults, who typically have serious problems perceiving speech in a noisy environment (Parbery-Clark et al., 2011)

THINKSTOCK

Remembering to be careful not to confuse correlation with causation, you may ask: Do differences between musicians and nonmusicians reflect the effects of musical training, or do they reflect individual differences, perhaps genetic ones, that lead some people to become musicians in the first place? Maybe people blessed with enhanced brain responses to musical or other auditory stimuli decide to become musicians because of their natural abilities. Recent experiments support a causal role for musical training. One study demonstrated structural brain differences in auditory and motor areas of elementary schoolchildren after 15 months of musical training (learning to play piano), compared with children who did not receive musical training (Hyde et al., 2009). Furthermore, brain changes in the trained group were associated with improvements in motor and auditory skills. Another study compared two groups of 8-year-old children, one group with 6 months of musical training and the other with 6 months of painting training. Musical training produced changes in the brain’s electrical responses to musical and speech stimuli, and those changes were correlated with enhanced performance on both musical and speech perception tasks (Moreno et al., 2009). And when students were randomly assigned to music or drama lessons, those who received music lessons showed an increase in IQ scores after training relative to those taking drama (Schellenberg, 2004). More recent musical training studies have also documented benefits for speech perception (François et al., 2013), and indicate that neural changes produced by musical training in childhood persist into adulthood (Skoe & Kraus, 2012).

We have yet to know all the reasons why musical training has such broad effects on auditory processing, but one likely contributor is that learning to play an instrument demands attention to precise details of sounds (Kraus & Chandrasekaran, 2010). Future studies will no doubt pinpoint additional factors, but the research to date leaves little room for doubt that your hours of practice were indeed worth the time.

  • Perceiving sound depends on three physical dimensions of a sound wave: The frequency of the sound wave determines the pitch; the amplitude determines the loudness; and differences in the complexity, or mix, of frequencies determines the sound quality or timbre.

  • Auditory pitch perception begins in the ear, which consists of an outer ear that funnels sound waves toward the middle ear, which in turn sends the vibrations to the inner ear, which contains the cochlea. Action potentials from the inner ear travel along an auditory pathway through the thalamus to the primary auditory cortex (area A1) in the temporal lobe.

  • Auditory perception depends on both a place code and a temporal code. Our ability to localize sound sources depends critically on the placement of our ears on opposite sides of the head.

  • Some hearing loss can be overcome with hearing aids that amplify sound. When hair cells are damaged, a cochlear implant is a possible solution.

164