54 | Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 Perspectives on the Emotional Response to Consonance and Dissonance in Music Thalia Vrantsidis Abstract In music, the tendency for certain combinations of notes to sound either pleasant (i.e., consonant) or unpleasant (i.e., dissonant) seems to be innate, universal, and uniquely human—provoking the question of why these intervals are able to trigger these emotional reactions. Evidence will be reviewed here to explore several aspects of this question: (1) which neural circuits are involved; (2) what acoustic properties lead to emotional reactions; and (3) why these preferences might have evolved. Brain imaging research shows that many emotion-processing areas respond differentially to consonant and dissonant music. In addition, lesion studies have found specialized neural circuitry involved in emotional responses to music. Studies of people with amusia, a pitch processing disorder, have identified the acoustic properties that determine a sound’s pleasantness—specifically, the harmonic series, rather than beating as previously assumed, distinguishes consonant from dissonant intervals. Consonant intervals are also the ones most prevalent in the harmonics of the human voice. This may explain why humans show a preference for consonant intervals. “Ah, music,” he said, wiping his eyes. “A magic beyond all we do here!” – Dumbledore, in J.K. Rowling’s Harry Potter and the Sorcerer’s Stone (1997, p. 95) Music has the power to make people feel. It inspires, it soothes, it makes us happy, and it makes us cry. Music is used to emotionally enrich the vast experiences of life, from milestone celebrations and sporting events, right down to personal listening (McDermott & Hauser, 2005). How is it that a pattern of notes can inspire such reactions? Part of the mystery of music is that there does not seem to be an obvious evolutionary benefit to it. Furthermore, it is not clear why certain patterns of notes produce certain emotional responses: why does Mozart sound beautiful, but a child smashing the keys of a piano sound terrible? Two of the most basic musical aspects that evoke emotional responses are consonance and dissonance, which define combina- tions of notes that sound either pleasant or unpleasant, respectively. There is widespread agreement that certain intervals sound highly pleasant, for example, a perfect 5th (defined by a distance of 7 semitones1), while others are regarded as highly unpleasant, especially the tritone (defined by a distance of 6 semitones) (Krumhansl, 1990; Malmberg, 1918; McDermott, Lehr, & Oxenham, 2010). This paper will examine the neuroscientific research to explore the neural underpinnings of the emotional reaction to consonance and dissonance, and discuss the acoustical properties of the sounds which trigger these responses. In addition, this research will be connected to a theory of why emotional responses to consonance and dissonance might have evolved. Preferences for consonant over dissonant intervals are a promising starting point for understanding human emotional reactions to music, as they appear to have an evolutionary basis that is unique to humans. Masataka (2006) showed that human infants as young Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 | 55 as two days old show preferences for consonant over dissonant music, indicating the preference is intrinsic in nature. This occurred equally in children of deaf and hearing parents, suggesting that these preferences are formed independently of pre-natal exposure to music, which presumably would differ with deaf and hearing parents. In addition, cross-cultural studies have shown similar preferences in many non-Western cultures (Butler & Daston, 1968; Fritz et al., 2009) although exceptions have been found in Indian listeners, perhaps due to the increased exposure to dissonance as used in Indian classical music (Maher, 1976). These results suggest that the human preference for consonance may be innate, although alterable by experience, and thus have an evolutionary basis. Furthermore, although non-human species such as birds and monkeys can be trained to differentiate between consonant and dissonant sounds (Hulse, Bernard, & Braaten, 1995; Izumi, 2000), humans are the only species known to show a preference between interval types (McDermott & Hauser, 2004). The emotional reactions to consonance and dissonance are an innate and uniquely human response, and can serve as a window into the human experience of music. Review of Neuroscientific Literature While the neuroscience of musical emotions is still in its infancy, it has made considerable strides in understanding the neural underpinnings of the emotional response to consonance and dissonance. Notably, it has been found that the brain regions which respond preferentially to either consonant or dissonant music are also involved in processing other pleasant or unpleasant emotional stimuli. For example, a study conducted by Blood, Zatorre, Bermudez, and Evans (1999) utilized positron emission tomography (PET) to investigate brain responses to music with varying degrees of dissonance. Dissonance was positively correlated with activation in the parahippocampal gyrus, an area previously shown to react to pictures with a negative emotional valence (Lane et al., 1997). Conversely, consonance was positively correlated with activation of the subgenual cingulate, frontopolar and orbitofrontal cortices. These regions have also been shown to be involved in emotional processing; for example, lesions of the subgenual cingulate and other ventral medial prefrontal cortex regions in monkeys have been shown to impair recognition of emotional expressions (Hornak, Rolls, & Wade, 1996). A study utilizing functional magnetic resonance imaging (fMRI) (Koelsch, Fritz, Müller, & Friederici, 2006) compared responses to joyful (i.e., mainly consonant) instrumental dance tunes and dissonant versions of those same tunes. The consonant music increased activation in the ventral striatum, containing the nucleus accumbens, which has been proposed as one of the main structures involved in the motivational and rewarding effects of positive stimuli (Cardinal, Parkinson, Hall, & Everitt, 2002; Salamone, Correa, Mingote, & Weber, 2005). The ventral striatum has also been shown to activate during the highly pleasurable experience of “chills” during music-listening (Blood & Zatorre, 2001). The dissonant music increased activation in the amygdala, hippocampus, parahippocampal gyrus, and temporal poles—areas that are highly interconnected and form a core network involved in reactions to aversive stimuli (see Koelsh et al., 2006, for a review). It is important to note that the involvement of the amygdala during emotional processing of music is still debated. It does not seem to be confined to aversive stimuli: Ball et al. (2007) showed no significant differences in amygdala activation between dissonant and consonant music, whereas Blood and Zatorre (2001) showed that the amygdala was involved during pleasurable musicinduced chills. Heart rate monitors and electroencephalography (EEG) have also been used to measure responses to consonant and dissonant music. Sammler, Grigutsch, Fritz, and Koelsch, (2007) found that, similar to unpleasant pictures, sounds, and films, dissonant music (relative to consonant music) led to a decreased heart rate, implying common underlying neural mechanisms 56 | Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 ing sound wave is experienced as having an unpleasant sound quality known as roughness. Beating was generally thought to be more prevalent in dissonant than consonant intervals and was thus used to explain the dislike of dissonant intervals (although see Cousineau, McDermott, & Peretz, 2012, for evidence that beating is generally equally prevalent in both types of intervals). A recent study tested this theory using people with amusia, a disorder characterised by difficulty in pitch processing (Cousineau, McDermott, & Peretz, 2012). Amusics showed no difference in liking of consonant and dissonant intervals; however, they still disliked notes with beating. Furthermore, McDermott, Lehr, and Oxenham (2010) found that, among individuals with normal auditory functioning, dislike of beating was not correlated with dislike of dissonant notes. These studies thus demonstrate that beating is not able to explain the preference of consonant over dissonant intervals. An alternative explanation exists based on the harmonic series, specifically the harmonics series present in the human voice. Naturally produced pitches do not contain a single frequency; instead, they are composed of a fundamental frequency accompanied by several sound waves with frequencies that are integer multiples of the fundamental frequency (i.e., harmonics; Cousineau, McDermott, & Peretz, 2012). Generally, the fundamental frequency is loudest (the wave has the largest amplitude), with each higher harmonic decreasing in volume (Smith, 2003). However, in ordinary human speech, certain harmonics tend to be louder, especially the third and fourth harmonics, due to properties Acoustical Basis of Consonance and Dissonance of the human vocal tract (Schwartz, Howe, & Purves, 2003). Schwartz et al. describe how this creates additional amplitude peaks, at certain frequencies particular Over the centuries, various theories have been to the human voice. They then show that the relationadvanced for what properties of consonant versus disship of the peaks to the fundamental frequency corresonant sounds might trigger the preference for consponds to the frequency ratios of each of the 12 notes sonant intervals (see Cousineau, McDermott, & Perof the chromatic scale compared to the tonic (the first etz, 2012, for a review). The most common theory is note). Furthermore, the relative amplitude of the peaks based on “beating”, a phenomenon in which two waves can be used to predict the consonance of each interval of different frequencies are combined, and its resultin the processing of these stimulus types. They also found an increase in frontal-midline (Fm) theta power for consonant music, indicating greater coherence of neural firing in certain population of neurons. This is interpreted as reflecting activation in the anterior cingulate cortex (ACC), which has been implicated elsewhere in processing pleasant music (Blood & Zatorre 2001). The ACC’s role in emotion is also supported by its inputs from other key emotion-related areas, such as the amygdala, hippocampus, and parahippocampal gyrus (Devinsky, Morrell, & Vogt, 1995). Despite the differences in brain activation patterns across these studies, it is clear that the major brain areas responsible for emotional processing are also being recruited for responding to consonant and dissonant music. In addition to these areas, Griffiths, Warren, Dean, and Howard (2004) may have discovered a brain area specific to musical emotions, rather than emotions in general. They reported a case study of a patient who, several months after a stroke, had selectively lost all emotional responses to music, while retaining his capacities for musical perception and emotional responses to non-musical stimuli. fMRI scans revealed damages mainly to the left insula, which also extended into the left frontal lobe and the left amygdala. This damage overlaps with areas shown to be active in normal emotional processing of music (Blood & Zatorre, 2001), suggesting that some of these damaged areas might form a specialized neural substrate for musical emotions, separate from music perception and nonmusical emotional responses. Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 | 57 in the chromatic scale, so that for larger peaks the corresponding interval is more consonant. Thus, the particular sounds that are considered consonant or dissonant seem to stem from the unique harmonic properties of the human voice. Based on these results, Schwartz and colleagues (2003) posited that the preference for consonance might have developed due to its biological utility. One of the primary tasks of the auditory system is to use sound information to identify biologically relevant stimuli. For humans, one of the most pervasive and biologically relevant sources of sound is human speech. However, the identity of a sound source is often ambiguous, so the brain relies on probabilistic rules to help identify the source—in this case, the fact that a certain pattern of harmonics is predictive of the human voice. Schwartz et al. thus proposed that a positive emotional reaction to the human voice might have generalized onto anything which resembled that harmonic pattern, leading to a preference for consonant intervals. This theory can explain the near universality of consonance rankings, as the same pattern of harmonics is present in all human speech, regardless of language (Schwartz et al., 2003). It also explains why other species do not show consonance preferences, since this harmonic pattern is unique to the human voice. Note that this theory does not specify whether the preferences are learned or are innate; however, the aforementioned developmental research would support an innate preference, for which Schwartz et al.’s theory could describe an evolutionary benefit. Conclusion Neuroimaging studies have broadened our understanding of the neural underpinnings involved in emotional responses to consonance and dissonance. These emotional responses generally involve the same neural circuitry that reacts to other pleasant and unpleasant stimuli. For consonance, this includes the subgenual and anterior cingulate cortex, the frontopolar and orbitofrontal cortex, the ventral striatum, and potentially the amygdala. For dissonance, it involves the parahippocampal gyrus, hippocampus, and temporal poles, and amygdala. The case study of a patient with lesions to the insula, amygdala, and frontal lobe implies that, within these regions, there is specific area for processing musical emotions in particular. Furthermore, studies of amusic individuals have revealed that the preference for consonance over dissonance does not, as previously thought, stem from the acoustical phenomenon of beating. Instead, statistical analyses suggest that it is specifically based on the similarity of consonant intervals to the harmonics of the human voice. This led Schwartz et al. (2003) to propose that a preference for the human voice provides a plausible explanation for the widespread and uniquely human preference for consonance over dissonance. Future Research One outstanding question which could be addressed by future studies is the degree to which innate emotional reactions to consonance and dissonance can be altered through experience, as was suggested by the finding that Indian listeners did not show the common pattern of consonance preferences (Maher, 1976). Future research could confirm this explanation by experimentally manipulating participants’ exposure to consonance and dissonance. Mull (1957) addressed this to some extent, showing that repeated exposure to dissonant music did in fact increase participants’ liking of it; however, it is possible that listeners did not actually enjoy the dissonance more, but were simply more able to focus on and appreciate the other aspects of the music. The use of more controlled stimuli along with neuroimaging could address this question more thoroughly. As a possible experiment, repeated listening over several days or weeks could be used to induce familiarity to either dissonant music or consonant music. Then single dissonant or consonant intervals could be played for the 58 | Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 Butler, J. W., and Daston, P. G. (1968). Musical consonance as musical preference: A cross cultural study. Journal of General Psychology, 79, 129–142. Cardinal RN, Parkinson JA, Hall J, Everitt BJ (2002). Emotion and motivation: the role of the amygdala, ventral striatum, and prefrontal cortex. Neuroscience & Biobehavioral Reviews, 26, 321–352. Cousineau, M., McDermott, J. H., & Peretz, I. (2012). The basis of musical consonance as revealed by congenital amusia. Proceedings of the National Academy of Sciences, 109, 19858–19863. Devinsky, O., Morrell, M. J., & Vogt, B. A. (1995). Contributions of anterior cingulate cortex to behaviour. Brain, 118, 279–306. Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., Friederici, A., & Koelsch, S. (2009). Universal recognition of three basic emotions in music. Current Biology, 19, 573–576. Griffiths, T. D., Warren, J. D., Dean, J. L., & Howard, D. (2004). “When the feeling’s gone”: a selective loss of musical emotion. Journal of Neurology, References Neurosurgery & Psychiatry, 75, 344–345. Ball, T., Rahm, B., Eickhoff, S. B., Schulze-Bonhage, A., Hornak, J., Rolls, E. T., & Wade, D. (1996). Face and voice expression identification in patients with Speck, O., & Mutschler, I. (2007). Response emotional and behavioural changes following properties of human amygdala subregions: ventral frontal lobe damage. Neuropsychologia, Evidence based on functional MRI combined with probabilistic anatomical maps. PLoS One, 34, 247–261. Hulse, S. H., Bernard, D. J., & Braaten, R. F. (1995). 2, e307. Auditory discrimination of chord-based Blood, A. J., & Zatorre, R. J. (2001). Intensely spectral structures by European starlings pleasurable responses to music correlate with (sturnus vulgaris). Journal of Experimental activity in brain regions implicated in reward Psychology: General, 124, 409–423. and emotion. Proceedings of the National Izumi, A. (2000). Japanese monkeys perceive sensory Academy of Sciences, 98, 11818–11823. consonance of chords. Journal of the Acoustical Blood, A. J., Zatorre, R. J., Bermudez, P., & Evans, A. C. Society of America, 108, 3073–3078. (1999). Emotional responses to pleasant and Koelsch, S., Fritz, T., Müller, K., & Friederici, A. D. unpleasant music correlate with activity in (2006). Investigating emotion with music: An paralimbic brain regions. Nature Neuroscience, fMRI study. Human Brain Mapping, 27, 2, 382–387. participants, rather than full songs, while they underwent fMRI and rated how much they liked the intervals. If familiarity with dissonant music did indeed increase participants’ liking of isolated dissonant intervals, this would be shown in the ratings; also, the brain scans for those exposed to dissonant music should show less activation in the areas associated with unpleasantness in previous studies. This study could thus demonstrate the extent to which the dislike of dissonance can be altered as a function of experience, and help to reconcile people’s innate preferences for consonance with the crosscultural variation found by Maher (1976). In addition to showing how the usual pattern of consonance preferences can be altered, these results could also highlight the potential importance of experience in strengthening the commonly-found consonance preferences, through a lifetime of exposure to consonant music and the human voice. Thus, although there does seem to be an innate basis for emotional reactions to consonance and dissonance, there is a role for both nature and nurture in a full explanation of this most basic aspect of the musical experience. Inkblot: The Undergraduate Journal of Psychology • Vol. 3 • September 2014 | 59 239–250. Krumhansl, C. L. (1990). Cognitive foundations of musical pitch (Vol. 17). New York, NY: Oxford University Press. Lane, R. D., Reiman, E. M., Bradley, M. M., Lang, P. J., Ahern, G. L., Davidson, R. J., & Schwartz, G. E. (1997). Neuroanatomical correlates of pleasant and unpleasant emotion. Neuropsychologia, 35, 1437–1444. Maher, T. F. (1976). “Need for resolution” ratings for harmonic musical intervals: A comparison between Indians and Canadians. Journal of Cross Cultural Psychology, 7, 259–276. Malmberg, C. F. (1918). The perception of consonance and dissonance. Psychological Monographs, 25, 93. Masataka, N. (2006). Preference for consonance over dissonance by hearing newborns of deaf parents and of hearing parents. Developmental Science, 9, 46–50. McDermott, J., & Hauser, M. D. (2004). Are consonant intervals music to their ears? Spontaneous acoustic preferences in a nonhuman primate. Cognition, 94, B11–B21. McDermott, J., & Hauser, M. (2005). The origins of music: Innateness, uniqueness, and evolution. Music Perception: An Interdisciplinary Journal, 23, 29–59. McDermott, J. H., Lehr, A. J., & Oxenham, A. J. (2010). Individual differences reveal the basis of consonance. Current Biology, 20, 1035-1041. Mull, H. K. (1957). The effect of repetition upon the enjoyment of modern music. The Journal of Psychology, 43, 155–162. Rowling, J.K. (1998). Harry Potter and the sorcerer’s stone. New York, NY: Scholastic. Salamone, J. D., Correa, M., Mingote, S. M., & Weber, S. M. (2005). Beyond the reward hypothesis: alternative functions of nucleus accumbens dopamine. Current Opinion in Pharmacology, 5, 34–41. Sammler, D., Grigutsch, M., Fritz, T., & Koelsch, S. (2007). Music and emotion: Electrophysiological correlates of the processing of pleasant and unpleasant music. Psychophysiology, 44, 293–304. Schwartz, D. A., Howe, C. Q., & Purves, D. (2003). The statistical structure of human speech sounds predicts musical universals. The Journal of Neuroscience, 23, 7160–7168. Smith, S. W. (2003). Fourier transform pairs. Digital signal processing: A practical guide for engineers and scientists. Burlington, MA: Elsevier Science. Footnotes A semitone is the smallest interval between notes used in ordinary musical scales (equivalent to the interval between a white key and an adjacent black key on a piano, or between a C and C#). 1
© Copyright 2025 Paperzz