Brain Products Press Release July 2012, Volume 43 User Research When processing the melody of speech is atypical: ERP insights into emotional prosody processing in Williams syndrome by Ana Patrícia Teixeira Pinheiro 1,2 1 Neuropsychophysiology Laboratory, School of Psychology, University of Minho, Braga, Portugal 2 Cognitive Neuroscience Laboratory, Department of Psychiatry, Harvard Medical School, Boston, USA Williams syndrome is a genetic neurodevelopmental disorder resulting from a submicroscopic deletion of approximately 1.6 Mb including 24-28 genes on the long arm of chromosome 7 (7q11.23) (Ewart et al., 1993). This syndrome is characterized by an uneven cognitive profile and intellectual disability (Martens, Wilson, & Reutens, 2008). In spite of early claims proposing the modular preservation of language in WS coexisting with severe deficits in visuospatial abilities, a large body of recent research has questioned these claims and demonstrated that language abilities of individuals with Williams syndrome are in line with their general cognitive abilities. Abnormalities have been shown in specific language subcomponents, including deficits in the recognition of emotion in speech stimuli, i.e. emotional prosody (Catterall et al., 2006; Plesa-Skwerer et al., 2006, 2007). Auditory emotional processing: the example of emotional prosody Of note, emotional prosody represents the non-verbal vocal expression of emotion (e.g. Wildgruber, Ackermann, Kreifelts & Ethofer, 2006). At the perceptual and physical levels, emotional prosody is conveyed primarily by means of pitch, intensity, duration and timbre (Juslin & Laukka, 2003; Scherer & Oshinsky, 1977; Schirmer & Kotz, 2006). Schirmer and Kotz argue that three main stages are involved in indexing emotional information from an acoustic signal. The first stage occurs around 100 ms, when the sensory Fig. 1: stages of emotional prosody processing of acoustic cues takes place. The second stage of emotional prosody processing occurs around 200 ms, corresponding to the extraction of emotional significance from acoustic cues (see also Paulmann & Kotz, 2008; Paulmann, Seifert, & Kotz, 2010). The third stage (around www.brainproducts.com 400 ms) is associated with higherorder cognitive processes that allow emotional significance to be Ana Pinheiro assigned to the output of sensory processes, including the integration of the emotional significance of acoustic cues with semantic information. Importantly, event-related potential (ERP) studies on emotional prosody processing have indicated the role of the P200 component as an index of the detection of emotional salience from speech stimuli, i.e. the second stage of emotional prosody processing (Paulmann & Kotz, 2008; Paulmann et al., 2010). However, a recent study demonstrated that the sensitivity to emotional salience can occur as early as 100 ms and is indexed by N100 amplitude modulated by the emotional valence of non-verbal vocalizations (Liu et al., 2012). Impaired assignment of emotional salience to a speech signal: the example of Williams syndrome The ability to successfully recognize emotional prosody in speech stimuli is one of the cornerstones of effective functioning in social environment. Even though the engaging style of conversation frequently reported in Williams syndrome (WS) suggests that social communication is a relative strength in this disorder (e.g. Gonçalves et al., 2010), behavioral studies on emotional prosody recognition in WS demonstrate that emotional prosody recognition is disrupted in WS when compared with typically developing controls (Catterall et al., 2006; Plesa-Skwerer et al., 2006, 2007). However, individuals with WS seem to perform better than participants with learning or intellectual disabilities on the recognition of emotional tone of voice in filtered speech (Plesa-Skwerer et al., 2006), suggesting that sensitivity to non-linguistic affective information may be less affected in WS. Notably, no event related potentials (ERP) studies of prosody processing have been conducted in WS. While reaction time measures and error data are indirect measures of all processes that took place before the response was made, ERPs probe the course of neurocognitive processes before a response is made or even in its absence. Together they make it possible to arrive at a more complete understanding of all cognitive processes involved and their putative abnormalities. It is not clear how lexical and supra-segmental features of speech signal may interact to convey emotion or how such processes differ from the processes of extracting emotional Brain Products Press Release information from a speech signal that carries suprasegmental information alone, in typically developing controls and WS. In our ERP study we examined the temporal course of emotional prosody processing in WS, and investigated the role of semantic information in different stages of vocal emotional processing in WS. The study design, subjects and results have been described in detail previously (Pinheiro et al., 2011), in Research in Developmental Disabilities. We hypothesized that all stages of vocal emotional processing would be impaired in WS, i.e. abnormalities in N100 and P200 components and reduced accuracy in recognition of emotional prosody. Twelve participants with WS (M ± SD = 17.3 ± 6.5 years) and a group of typically developing subjects individually matched for chronological age and gender were presented with neutral, positive and negative sentences in two conditions: (1) sentences with intelligible semantic information (‘semantic content condition’ – SCC); (2) sentences with unintelligible semantic information (‘pure prosody’ condition – PPC). Five of these participants (2 typically developing and 3 WS individuals) were excluded from final statistical analyses due to excessive artifacts. Fig. 2: SSC-PPS Stimuli were 216 auditory sentences (108 SCC sentences and 108 PPC sentences). In each sentence condition, 36 sentences had a happy intonation, 36 an angry intonation, and 36 a neutral intonation. Of note, in the PPC, the lexical-semantic information was absent due to the systematic distortion of the acoustic signal (as described in Pinheiro et al., 2011 and in Pinheiro et al., in press). However, even though all lexical www.brainproducts.com July 2012, Volume 43 and semantic information was suppressed by substitution of phones using Praat (http://www.fon.hum.uva.nl/praat/) and MBROLA (http://mambo.ucsc.edu/psl/mbrola/) software, the prosodic modulations of the original sentences were kept. In order to minimize cognitive demands, sentence conditions were not randomized within each experimental block. We first presented all sentences with intelligible semantic information (SCC), followed by all transformed sentences (PPC). We posited that switching between the two conditions in the same block could enhance cognitive demands, introducing a confounding variable in the results (as pointed out by Kotz et al., 2003) and in addition that switching between the two conditions within the same block could create a possible influence of normal speech on prosodic speech. After listening to each sentence, participants were asked to decide the emotional intonation underlying each sentence they have heard. While the participants listened to the sentences and decided the associated vocal emotion, the electroencephalogram (EEG) was recorded using QuickAmp EEG recording system (Brain Products, Munich, Germany), with 22 Ag-AgCl electrodes. Averages were computed using a 200-ms prestimulus baseline, timelocked to the sentence onset and spanning the length of a sentence (1500 ms after the onset of the sentence). This approach was adopted following all existing ERP studies of emotional prosody processing using non-spliced sentences (Paulmann & Kotz, 2008b; Paulmann et al., 2009) in order to assure the comparability of our results with previous studies. Although prosody is a supra-segmental feature of a speech signal not locked particularly to any given word, in the case of our design, the first word carried most of the prosodic information related to the emotional significance of an utterance. Time locking to the onset of the sentence, and thus to this first word, and extending the epoch across the entire sentence insured that all possible significant shifts in prosody and their underlying neural events were captured. After careful visual inspection of grand averages, three peaks were identified and selected for analysis: N100 (SSC: 100-200 ms; PPS: 100-160 ms), P200 (SSC: 200-320 ms; PPS: 160-260 ms) and N300 (SSC: 320-450 ms; PPS: 280380 ms). Different latency windows were selected for SSC and PPS after a careful inspection of grand average waveforms that indicated earlier ERP effects for PPS relative to SSC. Interestingly, in this study we observed a third ERP component (N300) that was not reported in the study of Paulmann and Kotz (2008) and Paulmann et al. (2010). This may be due to linguistic differences in the stimuli used in our experiment and in those previous studies (European Portuguese in the current study, and German in the study of Paulmann and Kotz and Paulmann et al.). The EEG data were analyzed using the software package Brain Analyzer Brain Products Press Release July 2012, Volume 43 1.05.0005 (Brain Products, Munich, Germany). al. (1993). Hemizygosity at the elastin locus in a developmental disorder, ERP results indicated group differences at the level of N100, P200, and N300, suggesting differential effects of sentence condition and emotion. When compared with typically developing controls, WS individuals showed reduced N100 for SCC sentences, indicating a differential sensory processing of speech carrying simultaneously semantic information (SCC) and speech with no intelligible semantic information (PPC), and suggesting that semantic information present in non-distorted speech signal may interact with early sensory processes making it more difficult for WS individuals to process ‘normal speech’ relative to ‘pure prosody’ speech signal. Williams syndrome. Nature Genetics, 5(1), 11-16. Also, more positive P200 was observed for SCC sentences, in particular for emotional intonations (happy and angry), in WS relative to typically developing controls, suggesting abnormal detection of emotional salience in vocal stimuli, and indicating that these abnormalities are dependent on the presence of semantic information. Gonçalves, O. F., Pinheiro, A. P., Sampaio, A., Sousa, N., Férnandez, M., & Henriques, M. (2010). The narrative profile in Williams Syndrome: There is more to storytelling than just telling a story. The British Journal of Developmental Disabilities, 56(111), 89-109. Juslin, P. N., & Laukka, P. (2003). Communication of emotions in vocal expression and music performance: different channels, same code? Psychol Bull, 129(5), 770-814. Liu, T., Pinheiro, A., Zhao, Z., Nestor, P. G., McCarley, R. W., & Niznikiewicz, M. A. (2012). Emotional cues during simultaneous face and voice processing: electrophysiological insights. PLoS One, 7(2), e31001. Martens, M. A., Wilson, S. J., & Reutens, D. C. (2008). Research Review: Williams syndrome: a critical review of the cognitive, behavioral, and neuroanatomical phenotype. Journal of Child Psychology and Psychiatry, 49(6), 576-608. Neville, H. J., Mills, D. L., Bellugi, U., & Hillsdale, N. J. E., 1994. (1994). The reduced N300 for both types of sentence conditions (SCC and PPC) in WS relative to typically developing controls also suggested abnormal processing of the emotional significance of the acoustic signal, irrespective of the semantic status of the sentences. Eff_ects of altered auditory sensitivity and age of language acquisition Finally, behavioral results demonstrated higher error rates in the WS group relative to typically developing controls but only for angry prosody. Paulmann, S., & Kotz, S. A. (2008). Early emotional prosody perception on the development of language-relevant neural systems: Preliminary studies of Williams syndrome. In S. Broman & J. Grafman (Eds.), Cognitive Deficits in Developmental Disorders: Implications for Brain Function (6783). Hillsdale, NJ: Erlbaum. based on different speaker voices. Neuroreport, 19(2), 209-213. Pinheiro, A. P., Galdo-Alvarez, S., Rauber, A., Sampaio, A., Niznikiewicz, M., & Goncalves, O. F. (2011). Abnormal processing of emotional prosody Conclusions This study was able to demonstrate that abnormalities in ERP measures of early auditory processing in WS (e.g. Neville et al., 1994; Pinheiro et al., 2010) are also present during the processing of emotional prosody information, spanning the three stages of vocal emotional processing (Schirmer & Kotz, 2006): extracting sensory information from the acoustic signal (N100), detection of emotionally salient acoustic cues (P200), and cognitive evaluation of the emotional significance of stimuli (N300 and error rates). Together, these findings suggest an interaction between disrupted lower-level sensory processes and higher-order cognitive processes in bringing about emotional prosody processing abnormalities in WS. They further suggest that the abnormal integration of perceptual and categorical information with the conceptual knowledge of negative emotion (angry prosody) may play an important role in inefficient negative emotion recognition. in Williams syndrome: an event-related potentials study. Research in Developmental Disabilities, 32(1), 133-147. Pinheiro, A. P., Galdo-Alvarez, S., Sampaio, A., Niznikiewicz, M., & Goncalves, O. F. (2010). Electrophysiological correlates of semantic processing in Williams syndrome. Research in Developmental Disabilities, 31(6), 1412-1425. Pinheiro, A.P., del Re, E., Mezin, J., Nestor, P.G., Rauber, A., McCarley, R.W., Goncalves, O.F., & Niznikiewicz, M. (in press). Sensory-based and higher-order operations contribute to abnormal emotional prosody in schizophrenia: an electrophysiological investigation. Psychological Medicine. Plesa-Skwerer, D., Faja, S., Schofield, C., Verbalis, A., & Tager-Flusberg, H. (2006). Perceiving facial and vocal expressions of emotion in individuals with Williams syndrome. American Journal of Mental Retardation, 111(1), 15-26. Plesa-Skwerer, D., Schofield, C., Verbalis, A., Faja, S., & Tager-Flusberg, H. (2007). Receptive prosody in adolescents and adults with Williams syndrome. Language and Cognitive Processes, 22(2), 247-271. Scherer, K. R., & Oshinsky, J. S. (1977). Cue utilization in emotion References: attribution from auditory stimuli. Motivation and Emotion, 1(4), Catterall, C., Howard, S., Stojanovik, V., Szczerbinski, M., & Wells, B. 331-346. (2006). Investigating prosodic ability in Williams syndrome. Clinical Schirmer, A., & Kotz, S. A. (2006). Beyond the right hemisphere: brain Linguistics and Phonetics, 20(7-8), 531-538. Ewart, A. K., Morris, C. A., Atkinson, D., Jin, W., Sternes, K., Spallone, P., et www.brainproducts.com mechanisms mediating vocal emotional processing. Trends in Cognitive Sciences, 10(1), 24-30.
© Copyright 2026 Paperzz