Neurobiological models of speech perception posit that both left and right posterior temporal brain regions are involved in the early auditory analysis of speech sounds. However, frank deficits in speech perception are not readily observed in individuals with right hemisphere damage. Instead, damage to the right hemisphere is often associated with impairments in vocal identity processing. Herein lies an apparent paradox: The mapping between acoustics and speech sound categories can vary substantially across talkers, so why might right hemisphere damage selectively impair vocal identity processing without obvious effects on speech perception? In this review, I attempt to clarify the role of the right hemisphere in speech perception through a careful consideration of its role in processing vocal identity. I review evidence showing that right posterior superior temporal, right anterior superior temporal, and right inferior / middle frontal regions all play distinct roles in vocal identity processing. In considering the implications of these findings for neurobiological accounts of speech perception, I argue that the recruitment of right posterior superior temporal cortex during speech perception may specifically reflect the process of conditioning phonetic identity on talker information. I suggest that the relative lack of involvement of other right hemisphere regions in speech perception may be because speech perception does not necessarily place a high burden on talker processing systems, and I argue that the extant literature hints at potential subclinical impairments in the speech perception abilities of individuals with right hemisphere damage.
A rich neuroscientific literature has established the importance of the brain’s left hemisphere for processing language. Early patient data demonstrated that damage to left superior temporal (Wernicke, 1874) and left inferior frontal (Broca, 1861) brain regions can lead to a loss of language abilities (i.e., aphasia), and recent studies also support a critical role for left hemisphere structures in the process of speech perception specifically. In particular, a wealth of neuroimaging evidence suggests that left superior temporal regions are important for imposing category structure on acoustically similar speech sounds (Desai et al., 2008; Liebenthal et al., 2010; Luthra, Guediche, et al., 2019; Mesgarani et al., 2014; Myers, 2007; Yi et al., 2019) and that left inferior frontal regions play a key role in differentiating between similar speech sound categories (Lee et al., 2012; Myers, 2007; Myers, Blumstein, et al., 2009; Rogers & Davis, 2018; Xie & Myers, 2018).
Relatively less is known about the extent to which the right hemisphere plays a role in speech perception, which may largely be a result of the fact that damage to the right hemisphere does not typically result in an aphasia (Blumstein & Myers, 2014; Turkeltaub & Branch Coslett, 2010). Instead, research on the right hemisphere’s role in language processing has largely focused on its high-level role in processing pragmatic information (Siegal et al., 1996) such as emotional prosody (Heilman et al., 1984), metaphorical language (Schmidt et al., 2007), and other forms of nonliteral language, including humor and sarcasm (Mitchell & Crow, 2005). While prominent neurobiological models (e.g., the Dual Stream Model; Hickok & Poeppel, 2000, 2004, 2007) have proposed at least some degree of right hemisphere involvement in processing phonetic information, the precise function of the right hemisphere in speech perception is relatively underspecified, especially compared to the more detailed characterization of the left hemisphere.
Notably, however, the right hemisphere has been heavily implicated in vocal identity processing—that is, in processing perceptual information about a voice in order to identify who is talking (Maguinness et al., 2018; Perrodin et al., 2015). Neuropsychological studies have linked right hemisphere strokes to deficits in identifying people by voice (Luzzi et al., 2018; Roswandowitz et al., 2018; Van Lancker & Canter, 1982; Van Lancker & Kreiman, 1987), though strikingly, patients with right hemisphere damage do not typically show frank deficits in speech perception. It is puzzling that these patients show deficits in vocal identity processing but not in speech perception, since talker processing and phonetic processing are known to be closely tied; the mapping between acoustic information and phonetic information can vary considerably across talkers, and theoretical accounts of speech perception argue that to perceive the speech signal accurately, listeners condition phonetic identity on talker information (Johnson, 2008; Joos, 1948; Kleinschmidt, 2019; Kleinschmidt & Jaeger, 2015). Given that phonetic processing is tightly linked to talker information, I suggest that by considering the role of the right hemisphere in processing nonlinguistic information about vocal identity, we might better understand the role of the right hemisphere in speech perception.
Note that in this review, I use the term “talker processing” largely to refer to the processing of voice information in support of processing speech, consistent with the use of the term “talker” in the speech perception literature. In contrast, I use “vocal identity processing” to refer to the processing of voice information to determine who is talking. These two processes are assumed to be theoretically distinct but to rely on some shared cognitive and neural architecture (Maguinness et al., 2018).
The structure of this review is as follows. After briefly discussing the interdependence between phonetic processing and talker processing, I review the existing literature on the role of the right hemisphere in vocal identity processing, paying careful attention to the contributions of different brain regions. I then consider current perspectives on the role of the right hemisphere in speech perception before closing with the hypothesis that the right hemisphere (and the right superior posterior temporal cortex in particular) may play an important role in allowing listeners to condition phonetic identity on talker information during speech perception.
How Is Phonetic Processing Linked to Talker Processing?
Individual talkers can differ substantially in how they produce their speech sounds, with talkers varying both In their use of rapid temporal cues such as voice-onset time (VOT; Allen et al., 2003) and in their use of spectral cues that indicate phoneme identity (Peterson & Barney, 1952). A vast literature indicates that listeners are highly sensitive to these talker-specific differences in phonetic variation and that they adjust the mapping between acoustic information and phonetic categories accordingly (e.g., Allen & Miller, 2004; Clayards et al., 2008; Kraljic & Samuel, 2005; Norris et al., 2003; Theodore & Monto, 2019). More generally, theoretical accounts of speech perception posit that listeners maintain distinct sets of beliefs about how different talkers produce their speech sounds (Kleinschmidt & Jaeger, 2015), meaning that phonetic processing is intrinsically linked to talker information.
The interdependence between phonetic processing and talker processing is further highlighted by studies showing that phonetic processing is facilitated when listeners are familiar with a particular talker (a talker familiarity effect) and by studies showing that talker processing is facilitated when listeners are familiar with the phonetic inventory of a particular language (a language familiarity effect). With regard to the former, several studies have found that talker familiarity leads to perceptual gains when processing speech in noise (Kreitewolf, Mathias, & von Kriegstein, 2017; Nygaard & Pisoni, 1998; Souza et al., 2013), and that talker familiarity makes it easier to selectively attend to one talker while ignoring another (Holmes et al., 2018; Holmes & Johnsrude, 2020; Johnsrude et al., 2013; Newman & Evers, 2007). With regard to the language familiarity effect, a number of studies have demonstrated that talker identification is facilitated when listeners hear speech in their native language (in which they are familiar with the phonetic category structure) compared to when they hear speech in a foreign language (in which they are not; Goggin et al., 1991; Perrachione & Wong, 2007). Talker familiarity effects can be understood by considering that when listeners receive practice with a particular talker, the acoustic dimensions that are relevant for processing that talker’s voice acquire distinctiveness; if the same dimensions are relevant for both talker processing and phonetic processing, then experience with a talker should incur performance benefits for phonetic processing (Nygaard & Pisoni, 1998). Similarly, language familiarity effects can be understood by recognizing that when listeners are familiar with the phonetic inventory of a particular language, the key acoustic-phonetic dimensions for that language likewise acquire distinctiveness—and if the same dimensions are relevant for talker processing, then experience with phonetic processing should yield benefits for talker processing. Taken together, such findings indicate that speech perception and talker processing are highly interrelated processes.
How Does the Right Hemisphere Support Vocal Identity Processing?
A focus on the right hemisphere regions involved in talker processing could inform neurobiological accounts of phonetic processing, at least to the extent that the same right hemisphere regions are recruited for both processes. The association between the right hemisphere and vocal identity processing dates back at least to early clinical studies by Van Lancker and colleagues, who demonstrated that right-hemisphere stroke patients were more likely than left-hemisphere patients to show impairments in identifying the voices of celebrities when performing a forced-choice task (Van Lancker & Canter, 1982; Van Lancker & Kreiman, 1987). Since then, neuroimaging studies have clarified the role of different right hemisphere regions in vocal identity processing (see Maguinness et al., 2018, for a recent review). As illustrated in Figure 1, these studies have revealed that vocal identity processing is largely supported by a set of temporal regions, with posterior temporal regions (shaded green in Figure 1) playing an important role in the early sensory analysis of vocal information, and anterior temporal regions (shaded blue) being important for vocal identity recognition. While not always recruited in vocal identity processing, right frontal brain regions (shaded pink) have been implicated in tasks that require listeners to make comparisons between voices, especially when comparing a vocal sample to a target voice.
Temporal lobe contributions to vocal identity processing
Neuroimaging evidence suggests that there is a posterior–anterior gradient in superior temporal lobe responses to vocal information, with right posterior temporal regions being thought to play a larger role in the general sensory processing of voice information (Andics, McQueen, Petersson, et al., 2010; Belin, Zatorre, Lafaille, et al., 2000; Schall et al., 2014; von Kriegstein & Giraud, 2004) and right anterior temporal regions being implicated in mapping from vocal information to a specific identity (Andics, McQueen, Petersson, et al., 2010; Belin, 2006; Belin, Fecteau, & Bédard, 2004; Imaizumi et al., 1997; Nakamura et al., 2001; von Kriegstein & Giraud, 2004). Support for the involvement of posterior superior temporal cortex in vocal identity processing comes from a wide range of studies, including a seminal fMRI study in which Belin, Zatorre, Lafaille, et al. (2000) examined cortical responses when subjects passively listened to human vocal stimuli (both speech sounds and nonspeech vocalizations like laughter) as well as to several types of control stimuli (such as animal sounds, bells, and speech-shaped white noise). Vocal stimuli elicited robust activation in the superior temporal sulcus (STS) bilaterally, but activation on the right was greater both in magnitude and in area than activation on the left. Notably, the response in the right STS was not specific to speech, as activation in the right posterior STS did not differ between speech and nonspeech human vocalizations. Belin, Zatorre, Lafaille, et al. further observed that band-pass filtering the stimuli led to a reduction of STS activation, and this reduction of activation was associated with worsened behavioral performance in a perceptual judgment task conducted outside the scanner (e.g., deciding whether the sounds were vocal or nonvocal). Such results indicate that the right STS is involved in differentiating between vocal and nonvocal auditory information but do not indicate whether it is necessary for such discrimination. Evidence for the latter comes from a study by Bestelmeyer et al. (2011). In that study, the authors first performed a functional localizer to identify the specific parts of right temporal cortex that were recruited when participants passively listened to voices compared to nonvocal auditory stimuli. Subsequent transcranial magnetic stimulation (TMS) to these regions impaired participants’ ability to discriminate between vocal and non-vocal sounds. Taken together, these findings suggest a critical role for the right posterior STS in processing the acoustic detail of human voices.
By contrast, more anterior regions in the right superior temporal cortex seem to be important when listeners need to map these acoustic details to a specific identity. Belin and Zatorre (2003) used fMRI to measure the habituation of neural regions in response to a train of stimuli presented over a short interval. The researchers found that the right anterior STS habituated (i.e., its activity diminished) when listeners encountered a stream of phonologically distinct syllables that were all spoken by the same talker. By contrast, this region did not habituate when listeners encountered a stream of phonologically identical syllables spoken by different talkers. In other words, this region’s response depended on who was producing the speech but not on what the content of the speech was. Convergent evidence comes from Formisano et al. (2008), who collected fMRI data while participants passively listened to different vowels spoken by different talkers. The authors then trained a machine learning algorithm to classify stimuli on the basis of talker identity (ignoring vowel identity) and found that the most discriminative voxels were located in right anterior STS. More recently, Luzzi et al. (2018) reported a case study of a patient who had suffered a stroke that affected his right anterior STS but did not affect posterior temporal regions; while the patient was unimpaired in his ability to indicate whether two voices were the same or different, he was no longer able to recognize his favorite singers on the basis of their voices alone. Overall, these findings suggest a role for right anterior temporal regions in recognizing vocal identity, as opposed to low-level processing of voice information.
Consistent with this view, a number of other studies have found that right anterior temporal regions are recruited when listeners must match vocal details to a known vocal identity. In an fMRI study by von Kriegstein, Eger, et al. (2003), for instance, greater right anterior STS activation was observed when listeners attended to vocal information compared to linguistic information. Similar results were observed in an MEG study by Schall et al. (2014), in which greater right anterior STS activity was observed when subjects had to match a sample of speech to a name compared to when they had to indicate whether a probe word had been present in the speech stream. Moreover, the authors observed a strong correlation between the degree of right anterior STS activity and subjects’ behavioral accuracy on this talker judgment task, suggesting that the variability in the activity of the right anterior STS might underlie individual differences in voice recognition. One way to conceptualize these results is to note that in both the study by von Kriegstein, Eger, et al. (2003) and the study by Schall et al. (2014), listeners were required to compare the incoming auditory signal to their internal representation of a particular vocal identity. As such, the findings indicate that the right anterior STS may play an important role in matching complex auditory objects to a stored vocal representation.
The suggestion that right anterior temporal regions are important for identifying a person on the basis of their voice is particularly striking given studies indicating that the right anterior temporal cortex is vital for person recognition more broadly (Gainotti, 2007). Individuals with damage to the right anterior temporal lobe may show selective impairments in identifying people on the basis of their faces (Damasio, 1990; Gainotti et al., 2003; Tranel et al., 1997) or voices (Gainotti et al., 2003) alone. As such, right temporal regions are thought to be critically involved in integrating perceptual information with conceptual person-specific knowledge (Gainotti, 2007). Consistent with this view, Ross et al. (2010) demonstrated that transcranial direct current stimulation of the right anterior temporal lobe modulated the likelihood that individuals would recover from a tip-of-the-tongue state when naming celebrities from their photographs. However, no such effect of stimulation was observed when subjects were shown photographs of famous places. Such findings point to a critical role of right anterior temporal regions in representing semantic knowledge about person identity specifically. As such, the involvement of right anterior temporal regions in vocal identity recognition may reflect access to multimodal information related to person identity (Maguinness et al., 2018; Perrodin et al., 2015).
While vocal identity processing is supported predominantly by right hemisphere regions, there has been some evidence for left hemisphere involvement in this process. In a study by von Kriegstein and Giraud (2004), for instance, listeners heard speech from talkers who were personally known to them, as well as speech from relatively unfamiliar talkers, to whom listeners’ previous exposure was limited to a few audio clips presented during a familiarization phase. Participants heard several sentences spoken by both the familiar and unfamiliar talkers; on each trial, they had to make a judgment either about the verbal content or about the vocal identity. Making judgments about vocal identity elicited robust activation of both the right posterior and right anterior STS, consistent with the characterization of the right posterior STS being involved in sensory processing of vocal identity and the right anterior STS being involved in vocal identity recognition. The researchers then examined whether functional connectivity with these right temporal regions differed as a function of whether the talkers were personally known to the participants. When participants listened to familiar talkers, there was robust connectivity among different subregions of the right superior temporal lobe. By contrast, when participants heard unfamiliar talkers, there was robust connectivity between the right posterior temporal lobe and the left posterior temporal lobe, suggesting that talker familiarity may modulate the involvement of left hemisphere regions in vocal identity processing. Other studies have supported the notion that the involvement of left temporal cortex in vocal identity processing may differ as a function of talker familiarity (Roswandowitz et al., 2018), and additional work suggests that language familiarity may similarly modulate the involvement of left hemisphere regions in vocal identity processing (Perrachione et al., 2009). Nonetheless, at least one study of stroke patients found that while individuals with right hemisphere damage were impaired in recognizing familiar voices, the performance of patients with left hemisphere damage was comparable to that of healthy controls (Lang et al., 2009); that is, there was no evidence for a left hemisphere role in processing familiar voices. Though additional work is needed to clarify the precise contributions of left temporal cortex, extant data suggest that left posterior temporal regions may play at least some role in vocal identity processing. Nevertheless, the role of the left hemisphere in processing vocal identity information is clearly limited, especially in contrast to the well-established role of the right hemisphere.
Frontal lobe contributions to vocal identity processing
In addition to a role for the right temporal lobe, some studies have posited a role for right frontal regions in vocal identity recognition, particularly during tasks that require listeners to categorize voices (Andics, McQueen, & Petersson, 2013; Jones et al., 2015; Zäske et al., 2017) or that require listeners to compare a voice sample to a referent in working memory (Stevens, 2004). Some evidence for the former comes from a study by Andics, McQueen, and Petersson (2013), who presented listeners with a vocal morph continuum where stimuli consisted of two different voices blended in different proportions. Training was used to establish a category boundary between the two voices, and participants then completed an fMRI session in which they had to categorize steps along the morph continuum. Subsequently, a second set of training sessions was administered to establish a new category boundary, after which participants completed a second fMRI session. The authors found that the activation of the right inferior frontal cortex depended on the proximity of a stimulus to the category boundary established during training (regardless of the precise acoustic details). These findings were interpreted as evidence that the right inferior frontal cortex supports the categorization of vocal stimuli into vocal identity categories, with the harder-to-categorize near-boundary stimuli eliciting more activation in right inferior frontal cortex. Consistent with this finding, Jones et al. (2015) observed that stroke patients who had damage to right frontal cortex were impaired in their categorization of talker gender when presented with stimuli from male–female continua; critically, the right STS was intact in these patients, suggesting that these results were not attributable to impairments in early sensory processing. Thus, the right inferior frontal cortex appears to play a critical role in allowing listeners to evaluate voices with respect to known vocal categories, whether these categories are task-relevant (e.g., ones established through training) or socio-indexically derived (i.e., categories based on talker-relevant social cues, such as gender or sexual orientation; Johnson, 2008; Munson, 2007).
The right frontal cortex has also been implicated in tasks that require listeners to compare one vocal sample to a second sample held in working memory. In an fMRI study, Stevens (2004) had participants listen to a series of stimuli while performing a two-back working memory task. On some blocks, they had to indicate whether the talker producing the current stimulus was the same as the talker who had produced the stimulus two items previously, and on other blocks, they had to indicate whether the same word had been produced two items previously. Subjects showed greater activation in the right middle frontal gyrus when performing the talker two-back task and greater activation in left inferior frontal gyrus when performing the word two-back task. Such a finding suggests a role for right frontal regions when subjects have to make explicit comparisons about vocal identity across stimuli.
Strikingly, the role of right frontal brain areas in vocal identity recognition seems to parallel a similar role for left frontal regions in phonological processing during speech perception. Just as the right inferior frontal cortex is strongly recruited when listeners hear stimuli near a vocal category boundary, the left inferior frontal cortex has been shown to be robustly activated by stimuli near a phonetic category boundary (Myers, 2007). Similarly, right frontal regions are recruited when demands on vocal working memory are high, just as left frontal regions are recruited when demands on phonological processing are high (Burton et al., 2000). More generally, the extant literature suggests that vocal identity processing is supported by a right-lateralized neural system, whereas speech perception is supported by an analogous left-lateralized system. To the extent that phonetic processing is influenced by talker information (as described in How Is Phonetic Processing Linked to Talker Processing?), it is worth considering how the right hemisphere may interact with the left to support speech perception; I turn to this question next.
How Might the Right Hemisphere Support Speech Perception?
Though the leftward lateralization of language processing represents a core feature of current neurobiological models of speech perception (Binder, Frost, et al., 1997; Binder, Swanson, et al., 1996; Geschwind, 1970; Hickok & Poeppel, 2000, 2004, 2007; Rauschecker & Scott, 2009), there is nevertheless some evidence that the right hemisphere—and right temporal cortex in particular—does play a role in speech perception. At least one study (Boatman et al., 1998) demonstrated intact syllable discrimination in a patient whose left hemisphere was sedated through a sodium amobarbital injection (Wada & Rasmussen, 1960), and functional neuroimaging studies of speech perception routinely implicate right temporal structures in speech perception (Belin, Zatorre, Hoge, et al., 1999; Blumstein et al., 2005; Davis et al., 2011; Giraud et al., 2004; Turkeltaub & Branch Coslett, 2010; Zatorre et al., 1996). More recently, a study by Kennedy-Higgins et al. (2020) found that listeners’ ability to repeat speech presented against background noise was impaired when they received TMS above either the left or right superior temporal gyrus (STG), but not when stimulation was performed at a control site. Collectively, such findings suggest a nonnegligible role for the right hemisphere in speech perception.
However, while left and right temporal structures are both routinely recruited for speech perception, they do not respond equally to acoustic information. In particular, left temporal regions seem to respond preferentially to rapid changes in the auditory signal, whereas right temporal regions appear to have a general preference for processing low-frequency modulations in the acoustic signal (Belin, McAdams, et al., 1998; Robin et al., 1990; Schwartz & Tallal, 1980; Scott et al., 2000). On the basis of these and other findings, Poeppel (2003) proposed the asymmetric sampling in time (AST) hypothesis. Under this view, the left hemisphere samples the speech signal at a relatively fast rate (40 Hz) and as such is well-suited for processing rapidly changing acoustic information (fluctuations on the order of approximately 25 ms); as such, left temporal processing is thought to be reflected in neuronal oscillations that occur in the gamma frequency band. By contrast, the right hemisphere has a slower rate of temporal integration (5 Hz), allowing it to process signal fluctuations that occur on the order of approximately 200 ms; right temporal activity is thought to be reflected in theta-band neuronal oscillations. Notably, the right hemisphere preference for low-frequency modulations has been observed both with speech (Abrams et al., 2008) and nonspeech stimuli (Boemio et al., 2005; Zatorre & Belin, 2001), suggesting that asymmetric sampling is a core property of temporal cortex rather than being specific to speech perception. Key to the AST hypothesis is the premise that the processing preferences of the two hemispheres depend on the physical properties of the auditory signal.
The AST can readily explain an association between the right hemisphere and processing the prosody of speech, for instance, as prosodic cues are conveyed over a relatively large temporal window (Poeppel, 2003). However, rightward lateralization is not always observed for prosodic processing, with the precise lateralization depending on a number of factors, including the control task used (Kreitewolf, Friederici, & von Kriegstein, 2014). Moreover, a number of studies have demonstrated left hemisphere involvement in prosodic processing when such information conveys linguistic information, whether lexical (Gandour, Tong, et al., 2004; Gandour, Wong, et al., 2002) or syntactic (van der Burght et al., 2019). In one such study, van der Burght et al. observed robust activation of the left inferior frontal gyrus when prosodic information in a speech sample determined syntactic structure but not when prosody was not needed for resolving the sentence’s syntax. These results are consistent with the view that while hemispheric asymmetries in processing auditory information may be partly attributable to the physical acoustic properties of the signal, the extent to which each hemisphere is involved may also largely depend on the functional use of the signal (Van Lancker, 1980).
The functional view predicts that right hemisphere involvement in speech perception is not limited simply to instances when listeners integrate auditory information over a long temporal window—rather, the involvement of the right hemisphere in speech perception may specifically reflect the process of conditioning phonetic processing on talker information (Kreitewolf, Gaudrain, & von Kriegstein, 2014b; Luthra, Correia, et al., 2020; Myers & Mesite, 2014; Myers & Theodore, 2017; von Kriegstein, Smith, et al., 2010). Some evidence for this hypothesis comes from a study by von Kriegstein, Smith, et al. (2010), in which listeners heard stimulus trains that varied in syllable identity, amplitude, and/or vocal tract length (an acoustic parameter that differs across talkers). Listeners performed either a one-back speech task (in which they had to indicate if the current stimulus matched the preceding stimulus in syllable identity) or a control task (either a one-back talker task or a one-back amplitude task). The authors observed that the left posterior STG was sensitive to vocal tract length (i.e., to acoustic information associated with talker identity). Moreover, von Kriegstein, Smith, et al. (2010) found that during the speech task, the functional connections between the left posterior STG and its right hemisphere analogue differed as a function of vocal tract length. The authors interpreted their findings as evidence that when listeners process talker-specific information in support of speech recognition, both the left and right temporal cortex are recruited.
Additional support for this perspective comes from a study by Myers and Theodore (2017), in which listeners were exposed to two talkers who differed in their productions of the sound /k/. Specifically, the talkers differed in whether they produced /k/ with a relatively short or long VOT (an acoustic-phonetic cue that distinguishes the voiceless sound /k/ from its voiced counterpart, /g/); notably, processing VOT requires integrating over a relatively short temporal window. After being familiarized with these two talkers, listeners completed an MRI scan during which they performed phonetic categorization on the words “cane” and “gain”; critically, during this phonetic categorization task, listeners heard both talker-typical and talker-atypical variants of the word “cane.” Myers and Theodore found that the functional activation of the right STG depended on whether the “cane” variant heard was typical or atypical of that talker. Such a result is consistent with the functional view of hemispheric asymmetries, which holds that despite being a short-duration cue, VOT would be processed by the right hemisphere if it was informative of talker identity. Additionally, the authors observed that the more typical the acoustic-phonetic variant was of a talker, the more tightly coupled the activity between the right STG and left temporal cortex. Taken together, these findings support the perspective that the right temporal cortex may support a listener’s ability to adapt to the idiosyncratic ways that different talkers produce their speech sounds; this may be achieved through the activity of the right temporal cortex itself or through interactions between the right temporal cortex and left temporal regions associated with phonetic processing.
While there are documented functional connections between left posterior temporal regions involved in phonetic processing and right posterior temporal regions involved in the early analysis of vocal detail, there does not appear to be a strong role for functional connections between left posterior temporal regions and other right hemisphere regions associated with vocal identity processing (Figure 2). In considering why this might be, it is worth noting that these other regions are primarily associated with explicitly mapping vocal information to a known identity (in the case of right anterior temporal areas) or are recruited only when listeners are tasked with categorizing or comparing between vocal samples (in the case of right frontal regions). That is, these regions are only recruited when demands on vocal identity processing are high.
In ecological instances of speech perception, however, listeners may not need to make explicit judgments about talker identity; indeed, listeners can typically leverage myriad sources of context to identify a talker’s intended phoneme, be they syntactic (Fox & Blumstein, 2016), semantic (Borsky et al., 1998), lexical (Ganong, 1980), or visual (Frost et al., 1988; McGurk & MacDonald, 1976). As such, the involvement of right anterior temporal and right frontal regions in phonetic processing may be limited to situations where the demands on the talker identification system is high, such that talker identity uniquely determines the mapping between acoustics and phonemes. I suggest that future studies assess this hypothesis directly, investigating both the activation of these right hemisphere regions and their functional connections to left temporal regions involved in phonetic processing.
Furthermore, the observation that naturalistic speech perception does not necessarily place a strong burden on talker processing systems may hint at why frank deficits in speech perception are not observed in individuals with right hemisphere damage. I suggest that the impact of right hemisphere damage (and damage to right posterior temporal cortex in particular) may only be observable in tasks that specifically require listeners to condition phonetic identity on talker information. Future work testing this hypothesis in right hemisphere patients will therefore be important in elucidating a potential subclinical impairment.
The acoustic signal simultaneously conveys linguistic information about speech sounds as well as nonlinguistic information about vocal identity, and in general, the process of speech perception is not independent from processing talker information (Mullennix & Pisoni, 1990). In this review, I have attempted to clarify the nature of right hemisphere involvement in speech perception by focusing on its role in vocal identity processing. As depicted in Figure 1, vocal identity processing entails the contributions of right posterior temporal cortex, right anterior temporal cortex, and right inferior/middle frontal cortex. Based on the functional view of hemispheric contributions to processing auditory information (Van Lancker, 1980), I presented evidence that the recruitment of right posterior temporal regions during speech perception may reflect the process of conditioning phonetic identity on talker information. I noted that right anterior temporal and right frontal regions are not strongly implicated during speech perception (Figure 2), and I suggested that the limited involvement of these regions may reflect the fact that in ecological speech perception, demands on talker processing are relatively low. In closing, I suggest that our understanding of the role of the right hemisphere in speech perception may be improved by focusing specifically on conditions where demands on talker processing are high (e.g., when a listener must appeal to talker information in order to know how to map the speech signal onto phonetic categories). Future work of this sort may also elucidate potential subclinical impairments in speech perception in individuals who have sustained damage to the right hemisphere.
Sahil Luthra, National Science Foundation (http://dx.doi.org/10.13039/501100008982), Award ID: Graduate Research Fellowship. James S. Magnuson, National Science Foundation (http://dx.doi.org/10.13039/501100008982), Award ID: NRT 1747486.
I am thankful to Emily Myers, Jim Magnuson, Rachel Theodore, Gerry Altmann, Eiling Yee, Jonathan Peelle, and three anonymous reviewers for their feedback on previous versions of this manuscript. This work was supported by an NSF Graduate Research Fellowship awarded to the author. The publication of this work was supported by the program in Science of Learning & Art of Communication at the University of Connecticut, which is supported by the National Science Foundation under Grant DGE-1747486.
Competing Interests: The author has declared that no competing interests exist.
Handling Editor: Jonathan Peelle