Is the processing of affective prosody influenced by spatial attention? an ERP study
© Gädeke et al.; licensee BioMed Central Ltd. 2013
Received: 26 September 2012
Accepted: 24 January 2013
Published: 29 January 2013
The present study asked whether the processing of affective prosody is modulated by spatial attention. Pseudo-words with a neutral, happy, threatening, and fearful prosody were presented at two spatial positions. Participants attended to one position in order to detect infrequent targets. Emotional prosody was task irrelevant. The electro-encephalogram (EEG) was recorded to assess processing differences as a function of spatial attention and emotional valence.
Event-related potentials (ERPs) differed as a function of emotional prosody both when attended and when unattended. While emotional prosody effects interacted with effects of spatial attention at early processing levels (< 200 ms), these effects were additive at later processing stages (> 200 ms).
Emotional prosody, therefore, seems to be partially processed outside the focus of spatial attention. Whereas at early sensory processing stages spatial attention modulates the degree of emotional voice processing as a function of emotional valence, emotional prosody is processed outside of the focus of spatial attention at later processing stages.
KeywordsVoices Prosody Spatial attention Emotion ERPs
Vocal prosody is one of the most important features of human communication enabling individuals to recognize the affective state of people in order to react quickly and appropriately in social situations. Changes in respiration, phonation, and articulation determine the acoustic signal of a voice and accompany emotional reactions, similar as changes in facial expressions .
The encoding of emotional prosody is based on various acoustic features, such as pitch and pitch variation, syllable duration, and voice quality . The quality of the voice is determined by different laryngeal and supralaryngeal aspects. The extraction of emotional prosody has been suggested to be automatic [3–5]. Event-related potential studies have shown that similar to the processing of facial expression, emotional vocalizations (such as screams) elicit a fronto-central positivity 150 ms after stimulus onset compared to frequency-matched artificial sounds . Other studies have focused on the dissociation between vocal emotional processing and the processing of person-identity information in human voices . Early ERP priming effects have been observed for happy voices but not for sad voices in the time range of the P2, a positivity with a latency of 200 ms, whereas the ERP speaker identity matching effect did not start until around 300 ms . The authors referred their results to physical differences related to emotional prosody. They suggested that higher frequencies are presented in happy voices than in sad voices, which might contribute to a faster and more efficient processing of happy vocal stimuli . Another study has shown a modulation of ERPs by the emotional valence of voices in the P200 time range as well . The authors interpreted their results as evidence for a rapid emotional decoding.
In humans, an enhanced brain activity has been observed to emotional compared to neutral voices in auditory association areas including the superior temporal sulcus (STS) [3, 8–10] and the right middle superior temporal gyrus [11–13]. These activation patterns have been observed irrespectively of which vocal prosody was used [3, 4, 9]. A recent study used functional near-infrared-spectroscopy (fNIRS) and reported an increased activation in the auditory cortex in response to pleasant and unpleasant sounds in comparison to neutral sounds . This activation pattern suggests that even sensory areas differentially respond to emotional prosodies, which nicely matches corresponding results in the visual system . However, brain imaging data do not allow to conclude whether changes of sensory cortex activity are due to changes in bottom up processing or due to feedback connections of higher cortical areas or the amygdala.
It has been a matter of debate whether the processing of emotional information depends on attentional resources . There are a number of studies suggesting that attention is necessary to select basic stimulus features in order to store relevant stimulus properties in working memory . However, there is evidence as well suggesting that emotional signals can be processed independent of attention and awareness and may guide attention to salient stimuli .
Although there are a few studies suggesting that the processing of facial expression requires attention [15, 18, 19] many of the recent studies are compatible with the view that emotional features can be processed automatically. For example, a processing of emotional faces has been observed outside the focus of attention  and even in the disregarded space by neglect patients [21, 22]. By contrast, additional studies have shown that attention is capable to further enhance the processing of emotional facial expressions [17, 23]. These results suggest for visual emotional stimuli both some attention independent processing but also some top down control. This combination seems highly efficient since many emotional stimuli in the environment are totally irrelevant for an individual. Thus, an individual must be able to inhibit an orienting to emotional stimuli in order not to interfere with current action goals.
Interestingly, brain networks involved in emotional processing show an overlap with attentional networks: Both attentional and emotional processes have been found to activate higher cortical areas such as parietal, frontal and cingulated areas as well as subcortical regions [17 for an overview]. However, the emotion specific activation of the amygdala might allow emotional information to be processed prior to the attentive stimulus selection  thus enabling emotional features to serve as exogenous cues that guide attention to relevant events.
While interactions between attention and emotional processing in the visual modality have been extensively studied, the question whether the processing of emotional prosody depends on attention or not has been addressed only recently. The situation for emotional voices might be quite different than for emotional faces, since the latter requires an orientation of the eyes toward visual stimuli in order to perceive them with a sufficient accuracy, while such an overt orienting response is not necessary in order to process emotional voices.
Sander and Scheich  found amygdala activity in response to affective non-verbal vocalizations (laughing and crying) regardless of whether the participants attended to the emotional valence of the stimuli or were engaged in a distracter task. Moreover, Grandjean et al.  showed that emotion-related activity in response to angry voices in the middle right STS did not vary with selective spatial attention, suggesting a preattentive processing of emotional prosody. Using the same paradigm, Sander et al.  replicated these findings and extended them to the amygdala. Additionally, as for visual neglect [21, 22], auditory extinction was found to be attenuated for stimuli with an emotional as compared to a neutral prosody . More recent studies, however, challenge the assumption of a total automaticity of emotional prosody processing. Mothes-Lasch et al.  presented voices of seven different emotional prosodies: Participants had to classify the gender of the speaker or they had to perform a difficult visual discrimination task. The authors found a higher response of the amygdala to angry compared to neutral voices only in the auditory but not in the visual task suggesting that orienting attention away from the auditory modality (intermodal attention) abolishes emotional prosody processing. By contrast, emotional voice can serve as an exogenous crossmodal attention cue. Brosch et al.  showed that reaction times in a dot probe task were shorter when the visual target was presented on the side at which an angry utterance (compared to a neutral utterance) was heard just before. The parallel recorded ERPs revealed an enhanced amplitude of the visual P1 to the target when an emotional voice was presented at the same side as the visual target compared to when the visual target was presented at the side of the preceding neutral voice .
Our study extends previous work by investigating the time course of spatial attention effects under different emotional prosody conditions. Thus, the present study complements the findings of imaging studies using a dichotic listening paradigm in order to analyze spatial attention effects on the processing of human angry and neutral voices .
The present ERP-study orthogonally manipulated the focus of spatial attention and the emotional prosody of the stimuli in order to analyze whether or not physically identical emotional stimuli are processed differently within and outside the focus of spatial attention. Pseudo-words comprising two identical (standards) or two different (deviants) syllables spoken by two different female voices were presented randomly on the left and on the right side of the participant. In different blocks, participants had to attend to one of the two spatial positions. They had to respond to infrequent (p=0.05) deviant stimuli presented at the attended position. The present experiment, therefore allowed for a direct comparison of the processing of physically identical emotional stimuli once when spatially attended and once when unattended.
Based on previous findings, we expected spatial attention to enhance ERPs to vocal stimuli starting around 100 ms after stimulus onset (i.e., the auditory N1).
If emotional prosody is processed in the absence of spatial attention, ERP modulations due to emotional valence are expected to be independent of the focus of spatial attention and thus additive to the ERP spatial attention effects. By contrast, if attention is necessary to process emotional valence, an effect of emotional valence is predicted only for the spatially attended stimuli or is different for spatially attended and unattended stimuli.
Seventeen healthy young student participants took part in the main experiment. According to self-report all participants had normal hearing and normal or corrected to normal vision. They were either paid for participation or received course credits. The experiment was conducted in accordance with the ethical guidelines laid down in the Declaration of Helsinki (2000). Because of low performance (see procedure), only data of thirteen participants (20 to 28 years, mean age 23 years, 7 females) were analyzed.
The final stimulus set consisted of nine different two-syllable pseudo-words (pronounceable German non-words) spoken by two actresses in four different emotional prosodies (neutral, happy, threatening and fearful), resulting in 72 physically different stimuli. Three of the pseudo-words were deviant stimuli (two different syllables); the remaining six were standards (two identical syllables).
Stimulus selection and evaluation
In a first step 48 two-syllable pseudo-words comprising two identical syllables (e.g. fefe, gigi), and 20 pseudo-words with two different syllables (the second consonant or vowel differed from the first; e.g. fefi, giki) were generated. Stimuli with two identical syllables were later used as standards, those with two different syllables as deviants or targets. Two actresses spoke these pseudo-words three times in four different emotional prosodies (happy, fearful, threatening, and neutral). Stimuli were recorded with a DAT recorder in an anechoic chamber. They were transferred to the computer and saved as wav-files. Preprocessing was done with the GOLDWAVE software (http://www.goldwave.com). The volume of the single sound files was equalized by setting the root mean square of each stimulus to 0.025.
The best two of the three recordings of each pseudo-word and each speaker were preselected by one of the authors (JG) for an evaluation study in which 24 students of the University of Marburg (19 to 34 years, mean 23 years, 22 females) rated each of the remaining 1088 stimuli (68 pseudo-words*2 voices* 2 versions* 4 emotions) on three dimensions (valence, dominance, arousal) using scales from −3 to +3 (valence: unpleasant (−3) – pleasant (3); dominance: submissive (−3) – dominant (+3); arousal: calming (−3) – stimulating (+3)).
The nine experimental stimuli (six standards and three deviants in the four emotional expressions spoken by the two voices) were selected using the criteria of duration as well as ratings of valence. Stimuli shorter than 250 ms or longer than 1020 ms were discarded. Rating values were transformed on a scale ranging from 1 (equivalent to −3) to 7 (equivalent to +3). In order to make sure that stimuli of different emotional categories would differ in their perceived valence, cutoff scores for the mean valence ratings were applied. These cutoff scores (neutral: 3.5 and 5.5; happy: 5 and 7; threatening: 1 and 2.5; fearful: 2 and 3.5) were defined to guarantee distinct stimuli for each voice in the four emotional categories.
For different sets of stimuli, analyses of variance with the factor Emotional Prosody (four levels) were calculated for the following dependent variables: Duration, Pitch, Intensity, Valence Rating, Dominance Rating, and Arousal Rating. For the final set of standard stimuli, the duration did not differ between emotion conditions (F(3,15) = 1.80, p > .1). Mean pitch differed between emotions (F(3,15) = 43.48, p < .001). The mean pitch of neutral vocal stimuli was significantly lower compared to happy, threatening and fearful voices (ps < .01 (neutral versus happy: t(5) = −8.53, p < .001; neutral versus threatening: t(5) = −11.63, p < .001; neutral versus fearful: t(5) = −7.40, p < .01). Moreover, happy voices had a higher pitch compared to threatening and fearful voices (ps < .05) (happy versus threatening; t(5) = 3.44, p < .05; happy versus fearful: t(5) = 4.99; p < .01). Threatening and fearful voices did not differ in the mean fundamental frequency (threatening versus fearful: t(5) = 0.392, p > .1).
Valence ratings (standard stimuli) depended on Emotional Prosody (F(3,15) = 308.76, p < .001). All pairwise comparisons between emotional prosodies were significant (all ps < .01). Dominance ratings depended on Emotional Prosody as well (F(3,15) = 667.27, p < .001). All differences between emotional prosodies in dominance ratings were significant (all ps < .01). By contrast, arousal ratings for the four emotional prosodies did not differ (F(3,15) = .15, p > .1).
Item statistics: Mean (M) and Standard error of the mean (SE) of duration, pitch, intensity, valence ratings, dominance ratings and arousal ratings of standard stimuli in the different emotional prosodies spoken by the two different voices
Intensity * (dB)
Valence rating (1–7)
Dominance rating (1–7)
Arousal rating (1–7)
Item statistics: Mean (M) and Standard error of the mean (SE) of duration, pitch, intensity, valence ratings, dominance ratings and arousal ratings of deviant stimuli in the different emotional prosodies spoken by the two different voices
Intensity * (dB)
Valence rating (1–7)
Dominance rating (1–7)
Arousal rating (1–7)
The difference in duration of the deviant stimuli in different emotional prosodies was not significant (F(3,6) = 7.661 p < .1). Dominance ratings depended on Emotional Prosody (F(3,6) = 409.98, p < .001). All pairwise comparisons were significant (all ps < .01), with the exception of the comparison neutral vs. happy (p > .1). By contrast, arousal ratings did not depend on Emotional Prosody (F(3,6) = .15, p > .1).
The final stimulus set consisted of nine different two-syllable pseudo-words (pronounceable German non-words) each spoken by the two actresses in four emotional prosodies, resulting in 72 physically different stimuli. Three of the pseudo-words were deviant stimuli (two different syllables), the remaining six were standards (two identical syllables).
In order to learn to discriminate the voices and to get familiar with the experimental procedure all participants took part in a 3–3.5 hours training session, one or two days prior to the EEG session. The training session consisted of five different units each of which had to be performed three times. The stimuli (six standards and two deviants in the four emotional expressions) were different from those used for the main experiment but came from the same stimulus pool. Since we did not analyze the factor voicea, the first four training units are not described in detail. The last phase of the training session was identical to an experimental block.
For the main experiment the stimuli were presented from two speakers positioned in front of the participant at a distance of 1.4 m, one 45° to the left and one 45° to the right of the participant. Stimuli of both voices and of all emotional prosodies were presented with an equal probability and in a random order from the left and right speaker. Stimulus onset asynchronies varied between 1300 ms to 1700 ms with a mean of 1500 ms.
Participants were instructed to attend to stimuli which were presented at one of the two spatial positions (left or right) and which were spoken by one of the two female speakers. Their task was to respond by lifting the left or right index finger out of a light gate whenever they detected one of the deviant stimuli spoken by the attended voice and presented at the attended position (i.e. targets). Response hand was switched after half of the trials (from left to right index finger or vice versa). A specific instruction concerning the varying affective prosodies was not given. Thus, there were four experimental conditions (attend voice I vs. attend voice II and attend left vs. attend right speaker). Only the spatial attention effects were analyzeda.
The experiment comprised 16 blocks lasting for six to seven minutes each (four blocks for each of the four experimental conditions). A block comprised 192 standard stimuli (80%) and 48 deviant stimuli (20%), 24 of which were targets (5%). Attention instruction was changed every two blocks. Participants were blindfolded throughout the experiment. The correct position of the cap was achieved by aligning it at the nasion, the inion, and the preauricular points. The participant’s head was immobilized by using a chin rest. Moreover, participants were instructed to avoid excessive blinking, eye and head movements during a run. Breaks after a block were allowed whenever the participant wanted. The duration of the EEG experiment without any breaks was about 1.5 hour. Including breaks, practice and the electrode-preparation and removal, the whole experimental session lasted between 5 and 6 hours.
The EEG was continuously recorded from 61 Ag/AgCl electrodes mounted equidistantly in an elastic cap (Falk Minow Services, Munich). The central electrode M_4 (see data analyses) is positioned between Fz and Cz of the international 10–20 system. The horizontal electro-occulogram (HEOG) was assessed with a bipolar recording of two electrodes attached to the outer canthi of the eyes, the vertical EOG (VEOG) was monitored with an electrode placed under the right eye against the common reference. All electrodes were referenced to the right earlobe during recording and were re-referenced off-line to the averaged left and right earlobe references. Impedances were kept below 5 kΩ for scalp recordings and below 10 kΩ for EOG recordings by preparing the skin of participants with Every (Meditec SRI, Negernbotel) and alcohol. ECI Electrogel (Electrocap International, Ohio, USA) served as the electrolyte for all electrodes. The ground electrode was placed on a position at the middle of the forehead (below Fpz). Signals were amplified with two SynAmps-amplifiers (NeuroScan, Inc. Sterling, USA). The sample rate was 500 Hz and the bandpass was set to 0.1 – 100 Hz. Signals were recorded continuously and saved on a hard disc.
A response was classified as a hit if it occurred within a time window of 200 to 1700 ms following a target stimulus. All other responses were considered as false alarms (FA). Trials in which the participants did not respond to a target were defined as misses. The miss rate was derived by dividing the total number of misses by the total number of target trials. Correct rejections were defined as non-responses to deviants at the non-attended location.
Performance accuracy for discriminating the positions as a function of emotional prosody was assessed for each participant by calculating d’ (d ’ = z(p(hit)) − z(p(FA)); ). The hit rate was defined as the number of correct responses to target stimuli spoken by the attended voice at the attended position divided by the total number of targets. The false alarm rate (FA rate) was defined as the number of incorrect responses to deviant stimuli spoken by the attended or by the unattended voice but at the unattended position divided by the number of deviants at the unattended position.
Mean reaction times (ms) and mean accuracy (%) for each emotional prosody (neutral, happy, threatening, fearful) with standard errors of the mean
Percent correct (%)
Reaction Times (ms)
The continuous EEG was epoched from 100 ms prior to the stimulus onset until 1000 ms after stimulus presentation separately for each participant and condition. The pre-stimulus interval was defined as baseline. Only segments following standard stimuli were analyzed, segments with responses to standard stimuli were discarded.
Significant contrasts (post-hoc t-tests, p < .05) between ERPs following stimuli in different emotional prosodies in the three time-windows (PA: position attended, PU: position unattended; I: ipsilateral to stimulation, C: contralateral to stimulation)
neutral – happy
Significant over both hemispheres
ipsilateral (I) / contralateral (C) effect
position attended (PA) / unattended (PU)
neutral – threatening
Significant over both hemispheres
ipsilateral (I) / contralateral (C) effect
position attended (PA) /unattended (PU)
Significant over both hemispheres
ipsilateral (I) / contralateral (C) effect
position attended (PA) / unattended (PU)
Significant over both hemispheres
1, 2, 3, 4, 5, 6
ipsilateral (I) / contralateral (C) effect
ipsilateral (I) / controlateral (C) effect
Significant over both hemispheres
ipsilateral (I) / contralateral (C) effect
position attended (PA) / unattended (PU)
threatening - fearful
Significant over both hemispheres
1, 2, 3, 4, 6, 7
ipsilateral (I) / contralateral (C) effect
position attended (PA) / unattended (PU)
For statistical analyses, mean amplitudes were calculated for the following three time epochs: first time window (110–150 ms), second time window (190–260 ms), and third time window (260–350 ms). Time windows were chosen by visual inspection and mean amplitudes were calculated around peak values of the grand average (N1, P2). We expected early attention and affective prosody effects in the time range of the N1 and P2.
For each time epoch mean amplitudes of ERPs were submitted to ANOVAs comprising the four repeated measurement factors, Spatial Attention (two levels: attended vs. unattended), Emotional Prosody (four levels: neutral, happy, threatening, fearful), Cluster (eight levels), and Hemisphere (two levels: ipsilateral vs. contralateral). Higher order interactions were followed up with appropriate sub-ANOVAs or t-tests. Recordings from the fronto-central electrode M_4 were analyzed without the factors Hemisphere and Cluster.
Additionally, we calculated ERP difference waves (attended minus unattended) for each emotion (neutral, happy, threatening, fearful). An ANOVA including the factors Emotional Prosody (neutral, happy, threatening, fearful), Hemisphere (contra versus ipsilateral) and Cluster (1–8) was run. The ANOVA for the electrode M_4 was run with the factor Emotional Prosody only.
All statistics were computed with the program package SPSS, subroutine GLM for repeated measurements. Greenhouse Geisser -corrected p-values are reported . In order to prevent an inflation of the alpha error, the Bonferroni-correction (corrected for six tests comparing the four emotional conditions) was applied.
In the following result section, we first report the behavioral data including d’ scores and IE scores. In the ERP result section we first report the results for site M_4 followed by the results for the analyses including all clusters.
IE scores (ms)
Using IE scores as dependent variable, the main effect of Emotional Prosody was marginally significant (F(3,36) = 3.49, p < .1). IE scores were lower for neutral (mean: 1563 ms, SE = 204) compared to threatening voices (mean: 2238 ms, SE = 283) (t(12) = −4.906, p < .01 (see Figure 2)). All the other comparisons were not significant (p > .1).
Event-related brain potentials
110–150 ms: In the latency range of the N1 the overall ANOVA revealed a main effect of Spatial Attention (F(1, 12) = 10.16, p < .01), a main effect of Emotional Prosody (F(3, 36) = 29.81, p < .001), and an interaction of Spatial Attention and Emotional Prosody (F(3, 36) = 3.50, p < .05). The N1 was more negative going to spatially attended than spatially unattended standards (Figure 3, difference waves in Figure 4). Separate subordinated ANOVAs for different emotional prosodies found a significant enhancement of the N1 to standard stimuli due to Spatial Attention for fearful stimuli (fearful stimuli: main effect Spatial Attention: F(1,12) = 19.92, p < .01), while this effect was not significant for neutral, happy, and threatening stimuli (neutral: F(1,12) = 2.84, p > .1; happy: F(1,12) = .65, p > .1; threatening: F(1,12) = .20, p > .1).
Moreover, subordinated ANOVAs confirmed that the effect of Emotional Prosody was significant at both the attended and the unattended location but the higher F-value for the attended (position attended: F(3,36) = 43.89, p < .001) than the unattended position (position unattended: F(3,36) = 5.55, p < .01) suggests a stronger Emotional Prosody effect at the attended location (Figure 5). Post-hoc t-tests revealed that ERPs to happy stimuli were significantly less negative going than ERPs to neutral stimuli at both attended and unattended positions (neutral vs. happy: attended: t(12) = −8.57, p < .01; unattended: t(12) = −3.78, p < .05). The ERP to threatening stimuli was less negative going than the ERP to neutral stimuli only if stimuli were attended (neutral vs. threatening: attended: t(12) = −5.50, p < .01; unattended t(12) = −2.33, p > .1). The ERP to fearful stimuli did not differ from the ERP to neutral stimuli, but the ERP to fearful stimuli was significantly more negative than the ERP to happy stimuli at the attended position (happy vs. fearful: attended: t(12) = 12.03, p < .01; unattended: t(12) = 2.71, p > .1). At the attended position ERP differences between threatening and fearful stimuli reached the significance level (t(12) = 8.47, p < .01) while the difference between happy and threatening voices was only marginally significant (happy vs. threatening: t(12) = 3.12, p < .1). Happy stimuli elicited the least negative going ERP, followed by threatening stimuli.
The ANOVA for the difference waves (ERP (attended) minus ERP (unattended), see Figure 4) found a significant main effect of Emotional Prosody (F(3, 36) = 3.50, p < .05)b.
190–260: The overall ANOVA revealed a highly significant main effect of Spatial Attention (F(1,12) = 13.40; p < .01) in the absence of a significant interaction of Spatial Attention and Emotional Prosody (F(3,36) =.37, p > .1). The Spatial Attention effect consisted of a more negative ERP to stimuli at attended positions (see Figures 3 and 4).
The ANOVA for the difference waves (attended minus unattended) did not reveal a main effect of Emotional Prosody (F(3,36) = .37, p > .1, see also Figure 4 and endnoteb).
260–350: The overall ANOVA revealed a highly significant main effect of Spatial Attention (F(1,12) = 36.18, p < .001) but no interaction of Spatial Attention and Emotional Prosody (F(3,36) = .07, p > .1). The Spatial Attention effect consisted of a more negative ERP to stimuli at attended positions (see Figures 3 and 4). The overall ANOVA found a significant main effect of Emotional Prosody (F(3,36) = 28.97, p < .001, see Figure 5). As revealed by post-hoc t-tests, the ERP to happy stimuli was less negative than the ERP to neutral voices (neutral vs. happy: t(12) = −5.82, p < .01), whereas the ERP to fearful stimuli was only marginally significant more negative than the ERP to neutral stimuli (neutral vs. fearful: t(12) = 2.80, p < .1). The ERP to threatening stimuli did not differ significantly from the ERP to neutral stimuli, but ERP-differences between happy and threatening (t(12) = 4.97, p < .01) as well as between threatening and fearful stimuli (t(12) = 3.97, p < .05) reached the significance level. Moreover, ERPs to happy and fearful stimuli differed from each other (happy vs. fearful: t(12) = 9.95, p < .01).
The ANOVA for the difference waves (attended minus unattended) did not reveal an effect of Emotional Prosody (F(3,36) = .07, p > .1, see Figure 4 and endnoteb).
Figure 6 shows the grand mean event-related potentials to voice stimuli with a neutral (a), a happy (b), a threatening (c) and a fearful (d) prosody.
Figure 1 (see methods section) summarizes the corresponding significant main effects of Spatial Attention and Emotional Prosody, and the interactions of these factors for the three time windows and for each cluster.
110–150 ms: In the overall ANOVA the interactions of Spatial Attention and Emotional Prosody (F(3,36) = 3.58, p < .05), of Spatial Attention and Cluster (F(7,84) = 3.51, p < .05) as well as of Spatial Attention, Emotional Prosody, and Cluster reached significance (F(21, 252) = 3.12, p < .05). The Spatial Attention effect was significant at cluster 3 (F(1, 12) = 7.18, p < .05), and marginally significant at clusters 1 and 4 (ps < .1). When analyzing emotional prosodies separately, a main effect of Spatial Attention was found for fearful stimuli (ANOVA for fearful stimuli only: main effect of Spatial Attention: F(1, 12) = 17.10; p < .01; interaction of Spatial Attention and Cluster: F(7, 84) = 7.41, p < .01); this effect was reliable at clusters 1, 3, 4, 6, and 7 (ps < .05)). An interaction of Spatial Attention and Cluster was significant for threatening stimuli (F(7, 84) = 3.35; p < .05) as well. However, no more than a marginally significant main effect of Spatial Attention was found for Cluster 7 (F(1, 12) = 3.62; p < .1).
Moreover, the overall ANOVA revealed a highly significant main effect of Emotional Prosody (F(3,36) = 27.72, p < .001), and an interaction of Emotional Prosody and Cluster (F(21,252) = 9.91, p < .001). Main effects of Emotional Prosody were highly significant at the attended position (F(3, 36) = 18.24, p < .001) as well as at the unattended position (F(3,36) = 12.22, p < .001). The Emotional Prosody * Cluster interaction was significant at both the attended and the unattended position (ps < .05). At clusters 1, 2, 3, 4, and 6, main effects of Emotional Prosody were confirmed for stimuli at the attended (all ps < .01) as well as for stimuli at the unattended position (all ps < .01). At the posterior clusters 7 and 8 the Emotional Prosody effect reached significance only at the unattended position (ps < .05). Table 4 summarizes the significant comparisons between emotional prosodies (posthoc t-tests) at the different clusters.
The ANOVA for the difference waves (attended minus unattended) revealed a significant effect of Emotional Prosody (F(3,36) = 3.58, p < .05; see endnoteb).
190–260 ms: More negative ERPs in the attended than in the unattended condition were observed. Because of significant interactions of Spatial Attention and Cluster (F(7,84) = 11.42, p < .01) as well as of Spatial Attention, Hemisphere and Cluster (F(7,84) = 6.53, p < .05) in the overall ANOVA, separate ANOVAs for single clusters were calculated (see Figure 1 (methods section)). The Spatial Attention effect was significant for all clusters (all ps < .05), and with the exception of cluster 7, the interaction of Spatial Attention and Hemisphere was significant for all clusters (all ps < .05) as well. At clusters 1, 3, 4, and 6, the main effect of Spatial Attention was found at contralateral as well as at ipsilateral clusters (all ps < .05). At clusters 2, 5, and 8, this effect was significant only at ipsilateral clusters (all ps < .05). Moreover, the overall ANOVA revealed a highly significant main effect of Emotional Prosody (F(3,36) = 23.44, p < .001), and a significant interaction of Emotional Prosody and Cluster (F(21, 252) = 14.16, p < .001). A significant main effect of Emotional Prosody was found at all 8 clusters (all ps < .05). Table 4 summarizes the significant comparisons between emotional prosodies (posthoc t-tests) at the different clusters.
The ANOVA for the difference waves (attended minus unattended) did not reveal a significant main effect of Emotional Prosody (F(3,36) = .432, p > .1; see endnoteb).
260–350 ms: A significant main effect of Spatial Attention was observed (F(1,12) = 7.99, p < .05). The interactions of Spatial Attention and Cluster (F(7,84) = 23.41, p < .001), of Spatial Attention and Hemisphere (F(1,12) = 11.12, p < .01), and of Spatial Attention, Hemisphere, and Cluster (F(7,84) = 4.85, p< .05) reached the significance level as well. The Spatial Attention effect was reliable at clusters 1 to 5 (all ps < .01) (Figure 1 in the methods section). At these clusters and at cluster 6, the Spatial Attention * Hemisphere interaction reached significance, (all ps < .05). At clusters 1, 3, and 4, the Spatial Attention effect was found at ipsilateral as well as at contralateral clusters (all ps < .01). At clusters 2 and 5, the effect was significant only at ipsilateral clusters (ps < .01), at contralateral clusters marginal effects were observed (ps < .1).
Moreover, in the overall ANOVA the main effect of Emotional Prosody (F(3,36) = 21.97, p < .001) as well as the Emotional Prosody by Cluster interaction were highly significant (F(21,252) = 20.42, p < .001). The effect of Emotional Prosody was significant at all clusters (all ps < .01) with the exception of the posterior cluster 8.
The ANOVA for the difference waves (attended minus unattended) did not reveal an effect of Emotional Prosody (F(3,36) = .22, p > .1; see also endnoteb).
In sum, the posthoc t-tests for single clusters largely confirmed the results obtained for M_4 (see Table 4 for a summary).
The present study investigated effects of spatial attention and emotional prosody on the processing of vocal stimuli. Two-syllable pseudo-words spoken in four emotional prosodies (neutral, happy, threatening, and fearful) were presented in a random order from two spatial positions. Participants attended to one position only in order to detect infrequent deviant stimuli. Both behavioral and ERP indices of stimulus processing were assessed. The main findings were as follows:
Even though marginally significant, neutral targets were detected with a higher precision than targets spoken in a happy, threatening or fearful prosody.
ERPs differed as a function of emotional prosody both at the spatially attended and spatially unattended location. Importantly, the early spatial attention effect (N1) was mostly pronounced for fearful stimuli. The N1 following fearful stimuli was more negative than the ERPs following neutral stimuli, while ERPs elicited by happy and threatening stimuli were less negative than ERPs to neutral stimuli. The Emotional Prosody effect was significant for stimuli at attended and unattended positions.
Behavioral processing of vocal prosody
In most previous studies, aversive stimuli have been found to be detected faster than neutral or positive emotional stimuli [33–36]. By contrast, in the present study, detection rates were higher and processing was more efficient for neutral compared to happy, threatening and fearful stimuli. In the present study targets were defined as two different rather than two identical syllables. If the emotional prosody is partially automatically extracted, as suggested by the present results, it seems plausible to assume that the emotional tone might have distracted the participants and thereby caused lower target identification based on syllables. However, this speculation has to be treated with caution since the number of targets was relatively low.
Early ERP modulations by spatial attention for fearful human voices
The finding of an enhanced negativity starting in the latency range of the N1 replicates once again the well known spatial attention effect discovered by Hillyard et al. . Early attention effects starting around 100 ms post-stimulus are generally found for easy to discriminate channels such as two locations one in the left and one in the right hemifield .
However, it might be wondered why reliable spatial attention effects starting in the N1 time range were mainly seen for fearful voices while later ERP spatial attention effects emerged similarly for all emotional prosodies. A specific processing for fearful stimuli has often been reported [39–41] and has been seen as adaptive in a social context.
On the other side, it has to be noticed that relatively long ISIs were used in the present study. It is well known that N1 attention effects are most likely if short ISIs are employed . This might explain why attention effects for most of the emotional voices were observed relatively late. However, on this background, the earlier emergence of spatial attention effects for fearful stimuli stress the preferred processing and specific enhancement of the processing of these stimuli by spatial attention.
According to Treue , bottom-up features of the stimulus itself and attentional top-down influences are integrated in a common saliency map which is a representation of the environment that weighs every input by its sensory features and behavioral relevance. Stimuli of high salience are processed even if unattended, while the processing of less salient unattended stimuli is suppressed. Applied to the present study, the spatial attention effect in the N1 time range might be interpreted as a release of an active suppression of the processing of task irrelevant fearful stimuli or as a further enhancing of the processing of this stimulus class. The finding that the N1 amplitude of ERPs to fearful stimuli arose as a function of spatial attention to the level of the N1 to neutral stimuli might be interpreted as evidence for the first interpretation.
It might be argued that emotion specific features (such as duration and low level acoustic characteristics) cause the differences between ERPs to emotional stimuli.
Of course, we cannot rule out the possibility that other low-level features contribute to the N1 Emotional Prosody effects. Accordingly, previous research has suggested that a number of acoustic features such as fundamental frequency (f0) and intensity differ among different emotional utterances [44, 45]. As described, our stimuli varied in fundamental frequency as a function of emotional category as well, suggesting that our actors were able to produce valid stimuli. Eliminating these features would be equal to eliminating emotional prosody. However, there are many arguments, why our results cannot be fully accounted by simple differences in physical stimulus features: First, we controlled for the duration of stimuli across all emotional prosodies. Second, the successful control of physical stimulus features is supported by similar latencies of the vertex potential across the four emotional prosodies (see Figure 5). Third, physical stimulus features would be expected to mainly affect exogenous ERPs, in other words maximally the vertex potential. However, we observed ERP differences as a function of Emotional Prosody for all analyzed time epochs. Fourth, ERP differences due to physical stimulus features are supposed to be independent of top-down modulation that is attention in our study. However, The N1 amplitude differences between different emotional prosodies were different for the attended and the unattended condition.
Modality specific processing of emotions and the influence of spatial attention
The present results coincide with the results reported by Sauter and Eimer  who used human vocalizations as stimuli and found an ERP positivity (150–180 ms) for different emotions (fear, disgust, achievement) compared with their spectrally rotated counterparts analog to previous findings for emotional faces  and pictures .
Similar findings for the processing of auditory and visual affective stimuli are surprising, however. For auditory processing, the existence of subcortical pathways to the amygdala has been demonstrated by studies of fear conditioning to acoustic stimuli in rats and guinea pigs [47, 48]. This subcortical route has been suggested to be particularly fast and to enable a quick detection of emotionally relevant stimuli. For the processing of facial stimuli, several studies have proposed a fast subcortical pathway as well [49–51]. However, according to Pessoa and Ungerleider  such a short route has not yet been unequivocally demonstrated in the visual system. Indeed, Pessoa and Ungerleider  have found an enhanced processing of emotional faces only when they were attended. These authors argue that a detailed analysis of facial features is rather impossible via a subcortical route (see  for a detailed discussion). This would argue for a quicker and more automatic processing of auditory emotional stimuli compared to visual emotional stimuli. On the other hand, the identification of the valence of auditory stimuli requires the integration of the auditory stream across an extended time epoch while visual scenes are picked up at one glance. However, in contrast to visual stimuli which in most cases need to be foveated in order to be identified, no overt shifts of attention are necessary to identify auditory stimuli. In line with this reasoning we found that ERPs differed as a function of emotional prosody both in the spatially attended and the spatially unattended channel. Thus, the emotional valence of auditory stimuli might be partially extracted automatically or at least without spatial attention. This conclusion is in agreement with the brainimaging results of Klinge et al. (2010)  who did not find a difference in amydala activation as a function of whether or not the emotional prosody of voices had to be attended. By contrast another fMRI study found that when all auditory input had to be completely discounted (intermodal attention)  a differential activation of the amydala for different emotional prosodies did not emerge. By manipulating spatial attention within the auditory modality, our experiment more resembles the within modality manipulation of Klinge et al. (2010) .
Therefore, we interpret our results in line with reports suggesting that the emotional valence of stimuli can be extracted in the absence of (at least spatial) attention, but that attention nevertheless modulates emotional processing [23, 54].
Finally, it has to be pointed out that we observed different prosody effects for early and later ERPs, i.e. processing steps. This observation might be related to the two roads of emotion processing proposed by LeDoux (cited from ). The “quick but dirty low road” is assumed to mediate an automatic processing of both attended and unattended events. While it seems not to be able to fully distinguish the whole range of emotional prosodies, e.g. two different aversive stimulus classes, such as fearful and threatening stimuli, at least not at unattended locations, the slower “high road” allows for a more elaborated and differential processing of both attended and unattended stimuli. This finding would suggest that later ERP effects of emotional prosody are independent of spatial attention.
The present results suggest that while emotional prosody is processed independent of spatial attention, spatial attention nevertheless modulates the degree of voice processing as a function of emotional valence at sensory processing stages. By contrast, at later stages emotional prosody is processed independent of the focus of spatial attention. Further research has to investigate whether this rule holds for stimuli of other modalities or other types of attention (such as intermodal attention, conscious vs. unconscious processing) as well.
aOriginally, the main experiment comprised an additional orthogonally manipulated factor. Participants had to selectively attend to one voice only. However, the voices of the two female speakers were too similar and participants did not manage to distinguish between them. Even after excluding participants (n = 4) with very low performance in discriminating the voices (d’ < 0.4), mean d’ (calculated as d’ = z(p(hit)) - z(p(FA)) (FA = False Alarms) for the remaining participants (n = 13) was low (d’ = .67; SE = .08).
bSee results of the overall ANOVA (The main effect Emotional Prosody in the ANOVA of the difference waves corresponds to the Spatial Attention * Emotional Prosody interaction in the overall ANOVA).
Superior temporal sulcus
- IE score:
Inverse Efficiency score
Analysis of Variance.
This study was supported by grant of the German Research Foundation (DFG, Ro1226/4-3) and European Research Council (ERC-2009-AdG 249425-CriticalBrainChanges) to BR. We would like to thank Regina Leitner and Manuela Weichenrieder for the registration of the stimuli, and Rachel Kludt and Katherine Medford for useful comments to the manuscript.
- Scherer KR: Vocal correlates of emotion. Handbook of Psychophysiology: Emotion and Social Behavior. Edited by: Manstead A, Wagner H. 1989, London: Wiley, 165-197.Google Scholar
- Spreckelmeyer KN, Kutas M, Urbach T, Altenmüller E, Münte TF: Neural processing of vocal emotion and identity. Brain Cogn. 2009, 69: 121-126. 10.1016/j.bandc.2008.06.003.PubMed CentralView ArticlePubMedGoogle Scholar
- Grandjean D, Sander D, Pourtois G, Schwartz S, Seghier ML, Scherer KR, Vuilleumier P: The voices of wrath: brain responses to angry prosody in meaningless speech. Nat Neurosci. 2005, 8: 145-146. 10.1038/nn1392.View ArticlePubMedGoogle Scholar
- Sander D, Grandjean D, Pourtois G, Schwartz S, Seghier ML, Scherer KR, Vuilleumier P: Emotion and attention interactions in social cognition: Brain regions involved in processing anger prosody. Neuroimage. 2005, 28: 848-858. 10.1016/j.neuroimage.2005.06.023.View ArticlePubMedGoogle Scholar
- Straube T, Mothes-Lasch M, Miltner WH: Neural mechanisms of the automatic processing of emotional information from faces and voices. Br J Psychol. 2011, 102 (4): 830-848. 10.1111/j.2044-8295.2011.02056.x.View ArticlePubMedGoogle Scholar
- Sauter DA, Eimer M: Rapid detection of emotion from human vocalizations. J Cogn Neurosci. 2010, 22: 474-481. 10.1162/jocn.2009.21215.View ArticlePubMedGoogle Scholar
- Paulmann S, Kotz SA: Emotional prosody perception based on different speaker voices. Neuroreport. 2007, 2: 209-213.Google Scholar
- Belin P, Zatorre RJ, Lafaille P, Ahad P, Pike B: Voice-selective areas in human auditory cortex. Nature. 2000, 403: 309-312. 10.1038/35002078.View ArticlePubMedGoogle Scholar
- Ethofer T, Anders S, Wiethoff S, Erb M, Herbert C, Saur R, Grodd W, Wildgruber D: Effects of prosodic emotional intensity on activation of associative auditory cortex. Neuroreport. 2006, 17: 249-253. 10.1097/01.wnr.0000199466.32036.5d.View ArticlePubMedGoogle Scholar
- Wildgruber D, Riecker A, Hertrich I, Erb M, Grodd W, Ethofer T, Ackermann H: Identification of emotional intonation evaluated by fMRI. Neuroimage. 2005, 24: 1233-1241. 10.1016/j.neuroimage.2004.10.034.View ArticlePubMedGoogle Scholar
- Ethofer T, Anders S, Erb M, Herbert C, Wiethoff S, Kissler J, Grodd W, Wildgruber D: Cerebral pathways in processing of affective prosody: a dynamic causal modeling study. Neuroimage. 2006, 30: 580-587. 10.1016/j.neuroimage.2005.09.059.View ArticlePubMedGoogle Scholar
- Ethofer T, Wiethoff S, Anders S, Kreifelts B, Grodd W, Wildgruber D: The voices of seduction: cross-gender effects in processing of erotic prosody. Soc Cogn Affect Neurosci. 2007, 2: 334-337. 10.1093/scan/nsm028.PubMed CentralView ArticlePubMedGoogle Scholar
- Mitchell RL, Elliott R, Barry M, Cruttenden A, Woodruff PW: The neural response to emotional prosody, as revealed by functional magnetic resonance imaging. Neuropsychologia. 2003, 41: 1410-1421. 10.1016/S0028-3932(03)00017-4.View ArticlePubMedGoogle Scholar
- Plichta MM, Gerdes AB, Alpers GW, Harnisch W, Brill S, Wieser MJ, Fallgatter AJ: Auditory cortex activation is modulated by emotion: a functional near-infrared spectroscopy (fNIRS) study. Neuroimage. 2011, 55: 1200-1207. 10.1016/j.neuroimage.2011.01.011.View ArticlePubMedGoogle Scholar
- Pessoa L: To what extent are emotional stimuli processed without attention and awareness?. Curr Opin Neurobiology. 2005, 15: 188-196. 10.1016/j.conb.2005.03.002.View ArticleGoogle Scholar
- Posner MI, Petersen S: The attention system of the human brain. Ann Rev Neurosci. 1990, 13: 25-42. 10.1146/annurev.ne.13.030190.000325.View ArticlePubMedGoogle Scholar
- Vuilleumier P, Armony J, Dolan R: Reciprocal links between emotion and attention. Human Brain Function. Edited by: Frackowiak RSJ. 2004, San Diego: Academic Press, 419-444. 2Google Scholar
- Eimer M, Holmes A, McGlone FP: The role of spatial attention in the processing of facial expression: an ERP study of rapid brain responses to six basic emotions. Cogn Affect Behav Neurosci. 2003, 3: 97-110. 10.3758/CABN.3.2.97.View ArticlePubMedGoogle Scholar
- Holmes A, Vuilleumier P, Eimer M: The processing of emotional facial expression is gated by spatial attention: evidence from event-related brain potentials. Cogn Brain Res. 2003, 16: 174-184. 10.1016/S0926-6410(02)00268-9.View ArticleGoogle Scholar
- Vuilleumier P, Armony JL, Driver J, Dolan RJ: Effects of attention and emotion on face processing in the human brain: an event-related fMRI study. Neuron. 2001, 30: 829-841. 10.1016/S0896-6273(01)00328-2.View ArticlePubMedGoogle Scholar
- Vuilleumier P, Schwartz S: Emotional facial expressions capture attention. Neurology. 2001, 56: 153-158. 10.1212/WNL.56.2.153.View ArticlePubMedGoogle Scholar
- Vuilleumier P, Armony JL, Clarke K, Husain M, Driver J, Dolan RJ: Neural response to emotional faces with and without awareness: event-related fMRI in a parietal patient with visual extinction and spatial neglect. Neuropsychologia. 2002, 40: 2156-2166. 10.1016/S0028-3932(02)00045-3.View ArticlePubMedGoogle Scholar
- McKenna M, Gutierrez E, Ungerleider L, Pessoa L: Attention increases selectivity to emotional stimuli. Neuroimage. 2001, 13: 443-10.1016/S1053-8119(01)91786-0.View ArticleGoogle Scholar
- Sander K, Scheich H: Auditory perception of laughing and crying activates human amygdala regardless of attentional state. Cogn Brain Res. 2001, 12: 181-198. 10.1016/S0926-6410(01)00045-3.View ArticleGoogle Scholar
- Grandjean D, Sander D, Lucas N, Scherer KR, Vuilleumier P: Effects of emotional prosody on auditory extinction for voices in patients with spatial neglect. Neuropsychologia. 2008, 46: 487-496. 10.1016/j.neuropsychologia.2007.08.025.View ArticlePubMedGoogle Scholar
- Mothes-Lasch M, Mentzel HJ, Miltner WH, Straube T: Visual attention modulates brain activation to angry voices. J Neurosci. 2011, 31: 9594-9598. 10.1523/JNEUROSCI.6665-10.2011.View ArticlePubMedGoogle Scholar
- Brosch T, Grandjean D, Sander D, Scherer KR: Behold the voice of wrath: Cross-modal modulation of visual attention by anger prosody. Cognition. 2008, 106: 1497-503. 10.1016/j.cognition.2007.05.011.View ArticlePubMedGoogle Scholar
- Brosch T, Grandjean D, Sander D, Scherer KR: Cross-modal emotional attention: emotional voices modulate early stages of visual processing. J Cogn Neurosci. 2009, 21: 1670-1679. 10.1162/jocn.2009.21110.View ArticlePubMedGoogle Scholar
- Green D, Swets J: Signal detection theory and psychophysics. 1966, New York: WileyGoogle Scholar
- Townsend JT, Ashby FG: Methods of modeling capacity in simple processing systems. Cognitive theory. Edited by: Castellan NJ, Restle F. 1978, Hillsdale, NJ: Erlbaum, 199-139. Volume 3Google Scholar
- Spence C, Kingstone A, Shore DI, Gazzaniga MS: Representation of visuotactile space in the split brain. Psychol Sci. 2001, 12: 90--93. 10.1111/1467-9280.00316.View ArticlePubMedGoogle Scholar
- Greenhouse SW, Geisser S: On methods in the analysis of profile data. Psychometrika. 1959, 24: 95-112. 10.1007/BF02289823.View ArticleGoogle Scholar
- Eastwood JD, Smilek D, Merikle PM: Differential attentional guidance by unattended faces expressing positive and negative emotion. Percept Psychophys. 2001, 63: 1004-1013. 10.3758/BF03194519.View ArticlePubMedGoogle Scholar
- Hansen CH, Hansen RD: Finding the face in the crowd: an anger superiority effect. J Pers Soc Psychol. 1988, 5: 917-924.View ArticleGoogle Scholar
- Öhmann A, Flykt A, Esteves F: Emotion drives attention: detecting the snake in the grass. J Exp Psychol G. 2001, 13: 466-478.View ArticleGoogle Scholar
- Öhmann A, Lundqvist D, Esteves F: The face in the crowd revisited: a threat advantage with schematic stimuli. J Pers Soc Psychol. 2001, 80: 381-396.View ArticleGoogle Scholar
- Hillyard SA, Hink RF, Schwent VL, Picton TW: Electrical signs of selective attention in the human brain. Science. 1973, 182: 177-180. 10.1126/science.182.4108.177.View ArticlePubMedGoogle Scholar
- Näätänen R: Processing negativity: an evoked-potential reflection of selective attention. Psychol Bull. 1982, 92: 605-640.View ArticlePubMedGoogle Scholar
- Eimer M, Holmes A: An ERP study on the time course of emotional face processing. Cogn Neurosci Neuropsychol. 2002, 13: 427-431.Google Scholar
- Pourtois G, Grandjean D, Sander D, Vuilleumier P: Electrophysiological correlates of rapid spatial orienting towards fearful faces. Cereb Cortex. 2004, 14: 619-633. 10.1093/cercor/bhh023.View ArticlePubMedGoogle Scholar
- Pourtois G, Dan ES, Grandjean D, Sander D, Vuilleumier P: Enhanced extrastriate visual response to bandpass spatial frequency filtered fearful faces: time course and topographic evoked-potentials mapping. Hum Brain Mapp. 2005, 26: 65-79. 10.1002/hbm.20130.View ArticlePubMedGoogle Scholar
- Hillyard SA, Hansen JC: Attention: Electrophysiological Approaches. Psychophysiology: Systems, Processes and Applications. Edited by: Coles MGH, Donchin E, Porges SW. 1986, New York: Guilford Press, 227-243.Google Scholar
- Treue S: Visual attention: the where, what, how and why of saliency. Curr Opin Neurobiol. 2003, 13: 428-432. 10.1016/S0959-4388(03)00105-3.View ArticlePubMedGoogle Scholar
- Patel S, Scherer KR, Sundberg J, Bjoerkner E: Acoustic markers of emotions based on voice physiology. Proceedings of the speech prosody. 2010, 100865: 1-4.Google Scholar
- Ladd D, Silverman K, Tolkmitt F, Bergmann G, Scherer KR: Evidence for the independent funcion of intonation contour type, voice quality, and F0 rance in signaling speaker affect. Journal of the Acoustical Society of America. 1985, 78: 435-444. 10.1121/1.392466.View ArticleGoogle Scholar
- Keil A, Bradley MM, Hauk O, Rockstroh B, Elbert T, Lang PJ: Large-scale neural correlates of affective picture processing. Psychophysiology. 2002, 39: 641-649. 10.1111/1469-8986.3950641.View ArticlePubMedGoogle Scholar
- LeDoux JE: In search of an emotional system in the brain: Leaping from fear to emotion and consciousness. The Cognitive Neurosciences. Edited by: Gazzaniga MS. 1995, Cambridge, MA: MIT Press, 1049-1061.Google Scholar
- Weinberger NM: Retuning the brain by fear conditioning. The Cognitive Neurosciences. Edited by: Gazzaniga MS. 1995, Cambridge, MA: The MIT Press, 1071-1089.Google Scholar
- De Gelder B, Pourtois G, van Raamsdonk M, Vroomen J, Weiskrantz L: Unseen stimuli modulate conscious visual experience: evidence from inter-hemispheric summation. Neuroreport. 2001, 12: 385-391. 10.1097/00001756-200102120-00040.View ArticlePubMedGoogle Scholar
- Morris JS, Öhmann A, Dolan RJ: A subcortical pathway to the right amygdala mediating “unseen” fear. Proc Nat Acad Sci. 1999, 96: 1680-1685. 10.1073/pnas.96.4.1680.PubMed CentralView ArticlePubMedGoogle Scholar
- Öhman A: Automaticity and the amygdala: nonconscious responses to emotional faces. Curr Dir Psychol Sci. 2002, 11: 62-66. 10.1111/1467-8721.00169.View ArticleGoogle Scholar
- Pessoa L, Ungerleider LG: Neuroimaging studies of attention and the processing of emotion-laden stimuli. Prog Brain Res. 2004, 2004 (144): 171-182.View ArticleGoogle Scholar
- Klinge C, Röder B, Büchel C: Increased amygdala activation to emotional auditory stimuli in the blind. Brain. 2010, 133: 1729-1736. 10.1093/brain/awq102.View ArticlePubMedGoogle Scholar
- Gur RC, Schroeder L, Turner T, McGrath C, Chan RM, Turetsky BI, Alsop D, Maldjian J, Gur RE: Brain activation during facial emotion processing. Neuroimage. 2002, 16: 651-662. 10.1006/nimg.2002.1097.View ArticlePubMedGoogle Scholar
- McGovern K: Emotion. Cognition, Brain, and Consciousness - An Introduction to Cognitive Neuroscience. Edited by: Baars BJ, Gage NM. 2007, : Elsevier / Academic Press, 369-389.Google Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.