Selective auditory attention, or selective hearing, is a process of the auditory system where an individual selects or focuses on certain stimuli for auditory information processing while other stimuli are disregarded.[1] This selection is very important as the processing and memory capabilities for humans have a limited capacity.[2] When people use selective hearing, noise from the surrounding environment is heard by the auditory system but only certain parts of the auditory information are chosen to be processed by the brain.
Most often, auditory attention is directed at things people are most interested in hearing.[3] Selective hearing is not a physiological disorder but rather it is the capability of most humans to block out sounds and noise. It is the notion of ignoring certain things in the surrounding environment.
In an article by Krans, Isbell, Giuliano, and Neville (2013), selective auditory attention can be seen through the process of the bottleneck effect, a process of the brain that inhibits processing of multiple stimuli. For example, a student is focused on a teacher giving a lesson and ignoring the sounds of classmates in a rowdy classroom (p. 53). As a result, the information given from the teacher is stored and encoded in the student's long term memory and the stimuli from the rowdy classroom is completely ignored as if it weren't present in the first place. A brain simply cannot for a sustained period collect all sensory information that is occurring in a chaotic real-world environment, so only the most relevant and important information is thoroughly processed by the brain.[4]
Early researches on selective auditory attention can be traced back to 1953, when Colin Cherry introduced the "cocktail party problem".[5] At the time, air traffic controllers at the control tower received messages from pilots through loudspeakers. Hearing mixed voices through a single loudspeaker made the task very difficult.[6] In Cherry's experiment, mimicking the problem faced by air traffic controllers, participants had to listen to two messages played simultaneously from one loudspeaker and repeat what they heard. This was later termed the dichotic listening task.[7]
Though introduced by Colin Cherry, Donald Broadbent is often regarded as the first to systematically apply dichotic listening tests in his research.[8] Broadbent used the method of dichotic listening to test how participants selectively attend to stimuli when overloaded with auditory stimuli; Broadbent used his findings to develop the filter model of attention in 1958.[9] Broadbent theorized that the human information processing system has a "bottleneck" due to limited capacity and that the brain performs an "early selection" before processing auditory information.[10] Broadbent proposed that auditory information enters an unlimited sensory buffer and that one stream of information is filtered out and passes through the bottleneck to be cohesive, while all others that are not selected quickly decay in salience and are not processed.[11] Broadbent's model contradicts with the cocktail party phenomenon because Broadbent's model predicts that people would never respond to their names from unattended sources since unattended information is discarded before being processed.
Deutsch & Deutsch's late selection model that was proposed in 1963 is a competing model to Broadbent's early selection model.[12] Deutsch & Deutsch's model theorizes that all information and sensory input are attended to and processed for meaning. Later in the processing routine, just before information enters the short-term memory, a filter analyzes the semantic characteristics of the information and lets stimuli containing relevant information pass through to short-term memory and removes irrelevant information. Deutsch & Deutsch's model for selective auditory attention suggests that weak response to unattended stimuli comes from an internal decision on informational relevance, where more important stimuli are prioritized to enter the working memory first.
In 1964, Anne Treisman, a graduate student of Broadbent, improved Broadent's theory and proposed her own attenuation model.[13] In Treisman's model, unattended information is attenuated, tuned down compared to attended information, but still processed. For example, imagine that you are exposed to three extraneous sources of sound in a coffee shop while ordering a drink (chatter, coffee brewer, music), Treisman's model indicates that you would still pick up on the latter three sounds while attending to the cashier, just that these extraneous sources of noise would be muffled as if their "volumes" were turned down. Treisman also suggests that a threshold mechanism exists in selective auditory attention in which words from the unattended stream of information can grab one's attention. Words of low threshold, higher level of meaning and importance, such as one's name and "watch out", redirects one's attention to where it is urgently required.
Selective auditory attention is a component of auditory attention, which also includes arousal, orienting response, and attention span. Examining selective auditory attention has been known to be easier in children and adults compared to infants due to the limited ability to use and understand verbal commands. As a result, most of the understanding of auditory selection in infants is derived from other research, such as speech and language perception and discrimination.[14] However, small amounts of selection in infants has been recorded with preference over an infant's mother's voice compared to another female,[15] one's native language over a foreign one,[16] and speech directed towards infants instead of speech in between adults.[17]
As through age, older children have an increased ability to detect and select auditory stimuli compared to their younger counterparts. This suggests that selective auditory attention is an age dependent ability that increases based on improvements in automatic processing of information.[18]
As children of lower ages demonstrate a lesser ability to detect and select auditory stimuli compared to their older counterparts, the ability to discriminate irrelevant information from relevant has shown to be lower in those of younger ages than in older ages. The ability to allocate attention to one message among interfering messages increases with age, particularly between the ages 5 through 12 and eventually evening out after that.
Factors that have shown to contribute to these heightened abilities include increased language ability and word familiarity as age increases.
Another factor could be that older children are more equipped to understand a task and the reward and/or punishment for being able to understand and complete a task, thus eliminate unnecessary stimuli more frequently. Using the incidental learning paradigm, it was measured that children ages 11 and up begin to be less likely to process incidental stimuli due to the development of strategies to actively process relevant information over irrelevant.[19]
All in all, the inability to filter out irrelevant information and/or allocate attention to relevant information leads back to developmentally immature attention allocation.[20]
In recent years, neuroimaging tools such as PET (Positron Emission Tomography) and fMRI (Functional Magnetic Resonance Imaging) have been very successful in neural operations with high spatial resolution. Specifically, fMRI has been used to find evidence for attention effects in the auditory cortex in multiple studies. Another study based on "classical" dichotic selective listening paradigms has been proven to be successful as well. The findings showed that the effects were larger in the cortex contralateral to the direction of attention[21] [22] [23] [24] and were interpreted as "selective tuning of the left or right auditory cortices according to the direction of attention"[24]
The prevalence of selective hearing has not been clearly researched yet. However, there are some that have argued that the proportion of selective hearing is particularly higher in males than females. Ida Zündorf, Hans-Otto Karnath and Jörg Lewald carried out a study in 2010 which investigated the advantages and abilities males have in the localization of auditory information.[25] A sound localization task centered on the cocktail party effect was utilized in their study. The male and female participants had to try to pick out sounds from a specific source, on top of other competing sounds from other sources. The results showed that the males had a better performance overall. Female participants found it more difficult to locate target sounds in a multiple-source environment. Zündorf et al. suggested that there may be sex differences in the attention processes that helped locate the target sound from a multiple-source auditory field. While men and women do have some differences when it comes to selective auditory hearing, they both struggle when presented with the challenge of multitasking, especially when tasks that are to be attempted concurrently are very similar in nature (Dittrich, and Stahl, 2012, p. 626).[26]
Selective hearing is not known to be a disorder of the physiological or psychological aspect. Under the World Health Organization (WHO), a hearing disorder happens when there is a complete loss of hearing in the ears. It means the loss of the ability to hear. Technically speaking, selective hearing is not "deafness" to a certain sound message. Rather, it is the selectivity of an individual to attend audibly to a sound message. The whole sound message is physically heard by the ear but the brain systematically filters out unwanted information to focus on relevant important portions of the message. Therefore, selective hearing should not be confused as a physiological hearing disorder.[27] Selective auditory attention is a normal sensory process of the brain, and there can be abnormalities related to this process in people with sensory processing disorders such as autism, attention deficit hyperactive disorder,[28] post traumatic stress disorder,[29] schizophrenia,[28] selective mutism,[30] and in stand-alone auditory processing disorders.[31]
Target speech hearing has been proposed for hearable devices like headsets and hearing aids to gives wearers the ability to hear a target person in a crowd.[32] [33] This technology use real-time neural networks to learn the voice characteristics of the target speaker, which is later used to focus on their speech while removing other speakers and noise.[34] [35] The deep learning-based device lets the wearer to look at the target speaker for three to five seconds to enroll them. The hearable device can then cancel all other sounds in the environment and play just the enrolled speaker’s voice in real time even as the listener moves around and no longer faces the speaker. This could benefit individuals with hearing loss as well as sensory processing disorders.