47 min read

In 1997, during a performance of Górecki's Symphony No. 3 in Carnegie Hall, a survey found that 38% of audience members experienced tears or extreme difficulty controlling tears at some point during the performance. Another 25% reported chills. The...

In This Chapter

Chapter 27: Emotion, Tension & Release — The Physics of Musical Feeling

Opening: The Problem of Musical Emotion

In 1997, during a performance of Górecki's Symphony No. 3 in Carnegie Hall, a survey found that 38% of audience members experienced tears or extreme difficulty controlling tears at some point during the performance. Another 25% reported chills. The symphony contains no text (until its third movement), no dramatic narrative, no extramusical program to explain the emotion. There is only music — patterns of sound organized in time — and a hall full of strangers weeping in synchrony.

This is the central puzzle of musical emotion. Music is, at the level of physics, pressure variations in air. At the level of psychoacoustics, it is patterns of pitch, rhythm, timbre, and dynamic. At the level of music theory, it is chord progressions, melodic lines, and formal structures. None of these descriptions, individually or together, seems to explain why it makes people cry. Yet it does, reliably, repeatedly, across cultures and across centuries. Understanding how it works — the mechanism by which organized sound generates felt emotion — is one of the deepest problems in the science of mind.

This chapter assembles the most complete account we currently have. It begins with the definitional problem (what kind of thing is musical emotion?), moves through the major theoretical frameworks, and then examines the specific physical features of music — harmonic tension, melodic motion, rhythmic placement — that generate emotional responses. Along the way, it confronts the fundamental philosophical question: is musical emotion reducible to the physics of sound plus the machinery of prediction, or is there something genuinely irreducible about felt musical experience?


27.1 What Is Musical Emotion? The Definitional Problem

Before asking how music generates emotion, we need to ask what kind of emotion music generates. This turns out to be more contested than it appears.

Three Positions on Musical Emotion

Music expresses emotions but doesn't cause them. On this view, associated with the philosopher Susanne Langer and, in a different form, with the musician-theorist Leonard Meyer's early work, music is representational: it has the form or contour of emotional experience without producing actual emotional states in listeners. A minor-key passage "expresses sadness" in the way that a sad painting expresses sadness — through resemblance — without necessarily making you feel sad. This view explains why people can enjoy sad music without wanting to feel sad.

Music causes genuine emotions. On this view, musical emotions are real emotions — physiologically and phenomenologically equivalent to emotions triggered by non-musical events. The racing heart during an exciting climax, the tears during a grief-saturated slow movement, the joy of a bright, fast major-key dance — these are not aesthetic approximations of emotion but the real thing, produced by a real stimulus. This view is supported by the physiological evidence: music reliably produces the autonomic signatures of emotion (heart rate changes, skin conductance, hormonal responses).

Music causes aesthetic emotions that are distinct from basic emotions. A third position — probably the most defensible — holds that music triggers a class of emotions that are related to but distinct from the basic emotions (fear, joy, anger, sadness, surprise, disgust). These "aesthetic emotions" include things like awe, wonder, nostalgia, the bittersweet feeling of beauty, the specific pleasure of surprise resolution — feelings that are characteristically triggered by art, ritual, and aesthetic experience and that are not straightforwardly reducible to basic emotional categories.

This third position is supported by work on what Dacher Keltner and Jonathan Haidt have called "elevation" — the emotion triggered by witnessing great moral or aesthetic achievement — and by research showing that people use different vocabulary to describe their responses to music than to describe their responses to ordinary emotional events.

💡 Key Insight: Felt vs. Perceived Emotion

Music neuroscience consistently distinguishes between felt emotion (the listener actually experiences an emotional state) and perceived emotion (the listener recognizes the emotional character of the music without personally experiencing that emotion). You can hear a piece of music as "angry" without feeling angry yourself. These two responses can be dissociated experimentally, involve partially different neural substrates, and respond differently to the acoustic features of the music. Most theories of musical emotion need to account for both.


27.2 The Three Theories of Musical Emotion

Three theoretical frameworks dominate the contemporary psychology of musical emotion. They are not mutually exclusive — they describe different aspects of a complex phenomenon — but they emphasize different mechanisms and make different predictions.

Meyer's Expectation Theory

Leonard Meyer's 1956 book Emotion and Meaning in Music proposed that musical emotion arises from the manipulation of musical expectation. A musical phrase sets up expectations — learned from one's cultural exposure to musical syntax — about what should come next. When those expectations are fulfilled, satisfied, or violated in interesting ways, emotion results.

For Meyer, emotion is the felt experience of expectation: tension is the state of wanting a resolution that hasn't arrived; release is the state of having arrived. Surprise is a violated expectation; satisfaction is a fulfilled one; longing is an expectation that keeps being deferred. The entire emotional arc of a musical work — from the first note to the final cadence — can be understood as a sequence of expectation-states being created, maintained, elaborated, and eventually resolved.

Meyer's framework is structuralist in spirit: it ties musical emotion to musical structure, specifically to the syntactic patterns that define what is "expected" in a particular musical tradition. It explains why musical sophistication matters — the more you know about musical syntax, the more refined your expectations, and the more nuanced your emotional responses.

Juslin's BRECVEMA Model

Patrick Juslin's BRECVEMA model (2013) is currently the most comprehensive taxonomic account of the mechanisms by which music generates emotion. The acronym stands for eight distinct mechanisms:

📊 Data/Formula Box: The BRECVEMA Mechanisms

Mechanism Description Timescale Learning Required?
Brainstem reflexes Loud, sudden, fast sounds trigger arousal via evolved startle response Milliseconds No
Rhythmic entrainment Motor system entrains to beat; induces arousal/valence via movement Seconds Minimal
Evaluative conditioning Music paired with positive/negative events takes on associated valence Variable Yes (implicit)
Contagion Acoustic features of music resemble emotional vocalizations; mirrored Seconds Minimal
Visual imagery Music induces internal visual images with associated emotional content Seconds Minimal
Episodic memory Music triggers autobiographical memories with original emotional context Seconds Yes (incidental)
Musical expectancy Violation and fulfillment of learned syntactic expectations Milliseconds Yes (cultural)
Aesthetic judgment Reflective evaluation of music's quality, craftsmanship, or beauty Seconds Yes

The BRECVEMA model's strength is its comprehensiveness: it acknowledges that musical emotion is not a single thing but a family of related processes, each with different timescales, different dependence on learning, and different neural substrates. Brainstem reflexes operate in milliseconds and require no cultural learning; aesthetic judgment operates over longer timescales and requires musical sophistication.

Huron's ITPRA Theory

David Huron's 2006 theory, detailed in his book Sweet Anticipation, extends Meyer's framework within an evolutionary/neuroscientific framework. ITPRA identifies five sequential phases in the response to any musical event:

Imagination: The brain generates a probabilistic prediction of what will come next, based on statistical regularities learned from musical exposure. This prediction is often partially conscious — the experienced listener can often "imagine" what the next note will be.

Tension: As the event approaches, the brain enters a preparatory state — increasing arousal, readying responses — based on the predicted outcome. This tension is the felt anticipation of a predicted event.

Prediction (response): At the moment the event occurs, the brain computes a prediction error — how different was the actual event from the predicted one?

Reaction: The immediate, automatic response to the event, which differs depending on whether the prediction was confirmed, violated in a positive direction, or violated in a negative direction.

Appraisal: The reflective, slower assessment of the event in its broader musical context — whether the violation was interesting, whether the resolution was satisfying, whether the event served the larger formal goals of the piece.

The ITPRA cycle repeats continuously throughout musical listening, operating simultaneously at multiple timescales (individual notes, beats, phrases, sections, movements). Musical emotion, on this account, is the stream of ITPRA responses — the felt sequence of imaginations, tensions, reactions, and appraisals that constitutes the experience of listening to music unfold in time.


27.3 Expectation as the Engine of Emotion: How Musical Tension Is Literally Predictive Tension

The concept of expectation is central to all three theories, but it is worth being precise about what musical expectation actually is and where it comes from.

Musical expectation is not a conscious, deliberate inference ("I predict that the next note will be E"). It is largely implicit and automatic — a rapid, pre-attentive probabilistic weighting of what should come next, generated by the brain's statistical learning system on the basis of prior exposure to musical patterns. Human infants can learn the statistical regularities of their ambient musical culture by 8–10 months of age, suggesting that musical expectation is built on the same implicit statistical learning machinery that underlies language acquisition.

The statistical structure of Western tonal music is well-characterized. In a C major context, the note G (the dominant) is by far the most probable continuation of most melodic patterns; the note B (the leading tone) creates intense expectation for C; the tonic triad (C-E-G) provides the most stable resting point. These probabilities are not arbitrary — they reflect centuries of cultural selection for patterns that generate interesting, non-random sequences of expectation and resolution.

💡 Key Insight: Information Theory and Musical Tension

From the perspective of information theory, a fully expected event carries zero information — it adds nothing to the listener's knowledge. A maximally unexpected event carries maximum information — but if the music is entirely random, there is no expectation system to violate and no emotional response. Musical emotion emerges from the middle zone: events that are neither fully predictable nor completely random, where prediction is possible but not certain. This is sometimes called the "sweet spot" of musical complexity — and it explains why both extremely repetitive music and extremely avant-garde music can feel emotionally flat.


27.4 Tension and Release: A Physical Account

The subjective experience of musical tension and release has direct physical correlates in the acoustic signal. Understanding these correlates allows us to analyze musical emotion as a physical phenomenon without abandoning the reality of the subjective experience.

Harmonic Tension: Dissonance to Consonance

A chord's degree of harmonic tension is related to its acoustic roughness — the extent to which its component frequencies produce beats (amplitude fluctuations) with each other. Two frequency components produce beats when they are close enough in frequency that the auditory system cannot fully resolve them; the beats produce a rapid fluctuation in loudness that is perceived as roughness or dissonance.

In the chromatic scale, intervals can be ordered from most consonant to most dissonant: - Octave (2:1): essentially no beats; very consonant - Perfect fifth (3:2): few beats; very consonant - Perfect fourth (4:3): few beats; consonant - Major third (5:4): some beats; mildly consonant - Minor third (6:5): slightly more beats; mildly consonant - Tritone (√2:1 in equal temperament): many beats; dissonant

In harmonic progressions, chords with more dissonance create tension; chords with more consonance create resolution. The experience of "wanting" a dissonant chord to resolve to a consonant one is the musical experience of physical tension seeking physical resolution.

Melodic Tension: High vs. Low

Melodic tension is associated with several physical parameters:

Pitch height: Higher pitches are generally experienced as more tense than lower pitches. This may relate to acoustic properties (higher partials create more beating with lower notes), to evolutionary associations (high-pitched sounds are often associated with alarm calls), or to embodied metaphor (high/low as a spatial dimension of effort and release). In Western tonal music, the leading tone — the seventh scale degree, just a half-step below the tonic — creates intense melodic tension by virtue of its high pitch relationship to the expected resolution.

Melodic motion direction: Upward melodic motion typically increases tension; downward motion typically releases it. This is a physical correlate of the universal vocal acoustics of emotional arousal: emotional arousal raises fundamental frequency; emotional relaxation lowers it.

Melodic leap vs. step: Large leaps create more tension than stepwise motion (which is more predictable). After a large upward leap, there is statistical pressure (and perceptual expectation) for the melody to descend by step.

Rhythmic Tension: Off-Beat to On-Beat

Rhythmic tension is created by displacement from the strong metric positions. A note that falls on a weak beat — or, more extremely, between beats (syncopation) — is experienced as rhythmically tense because it violates the predictive pulse of the meter. Resolution occurs when the off-beat motion is followed by an on-beat arrival.

Jazz improvisation constantly exploits rhythmic tension: a skilled improviser will deliberately "sit behind the beat" or anticipate it, creating rhythmic tension that is resolved when the next strong beat arrives. This rhythmic tension-release cycle operates within the framework of the predictive pulse, and it is one of the primary mechanisms by which jazz creates forward momentum and emotional engagement.

⚠️ Common Misconception: Tension Is Unpleasant

The correspondence between musical tension and acoustic dissonance/rhythmic displacement can suggest that tension is simply unpleasant and resolution simply pleasant. This is wrong in two ways. First, listeners often report that highly dissonant, tense music is exhilarating rather than unpleasant — the tension has pleasurable qualities of its own, particularly when it is clearly directed toward an anticipated resolution. Second, resolution without preceding tension is bland — the most satisfying resolutions follow periods of prolonged tension. Musical pleasure is the arc from tension to release, not simply the endpoint.


27.5 The Physics of a Cadence: Why the Dominant-to-Tonic Movement Feels Like "Arriving Home"

The cadence — the harmonic progression that ends a musical phrase or section — is the most fundamental unit of musical tension-resolution in Western tonal music. The most conclusive is the authentic cadence: a dominant (V) chord resolving to a tonic (I) chord. In C major, this is a G major chord (G-B-D) resolving to a C major chord (C-E-G).

Why does this feel like "arriving home"?

The Acoustic Account

The dominant chord in just intonation contains a tritone — the interval between the third (B) and the seventh (F) of the dominant seventh chord (G-B-D-F). The tritone is the most dissonant interval in Western music, with two strong component frequencies close enough to produce significant beating. When the dominant seventh resolves to the tonic, this tritone resolves: the B moves up by a half-step to C (resolving its tension by reaching the tonic root), and the F moves down by a half-step to E (resolving to the tonic third). The acoustic roughness of the tritone is replaced by the smoothness of the major third.

The Statistical Account

In tonal music, the dominant-to-tonic progression is the most common phrase-ending progression. Listeners trained in Western music have heard this progression tens of thousands of times and have built an extremely strong statistical expectation for it. When it occurs, it confirms one of the most deeply ingrained musical expectations, triggering the reward of confirmation.

The Embodied Account

The dominant-to-tonic resolution may also have embodied correlates: a physical sensation of arrival, of weight settling, of effort completed. Many listeners report the experience of "landing" or "returning" at a tonic cadence — sensations that map onto physical experiences of movement and arrival. Whether this embodied quality is cause or effect of the emotional response is debated.

💡 Key Insight: Cadential Hierarchy

Not all cadences are equal. The perfect authentic cadence (V→I, both chords in root position, melody ending on the tonic) is maximally conclusive. The half cadence (ending on V) is maximally tensing — it creates a half-cadence "cliff" that demands continuation. The plagal cadence (IV→I, the "Amen" cadence of church music) is conclusive but softer. The Deceptive Cadence (V→vi, subverting the expected resolution) is the subject of the next section. These distinctions map onto physically distinct acoustic events, not arbitrary cultural conventions.


27.6 The Deceptive Cadence: Expectation Betrayed

The deceptive cadence — V→vi rather than V→I — is one of the most effective single-moment emotional devices in tonal music. When the dominant chord has built maximum expectation for tonic resolution, the arrival of the submediant (vi) instead triggers an immediate, powerful response: surprise, a slightly uncomfortable feeling of not-quite-arriving, followed immediately by a fresh charge of forward momentum.

The physics is precise. The dominant seventh chord (G-B-D-F in C major) sets up strong statistical expectation for C major (C-E-G). The A minor chord (A-C-E) that arrives instead shares two notes with the expected C major chord — C and E — but has replaced the expected tonic (C) in the bass with A. This partial confirmation (two shared tones) combined with partial violation (unexpected bass) creates a specific emotional quality: something like resolution-but-not-quite, arrival-without-rest.

The brain's ITPRA cycle explains the phenomenology: the tension phase is normal (strong expectation for tonic); the prediction moment arrives; the reaction is surprise (positive prediction error); the appraisal reassesses the new harmonic context. The next forward momentum begins from this new platform.

Great composers use the deceptive cadence to extend and intensify sections: by subverting the expected resolution, they create more tension, more forward drive, and ultimately make the real resolution, when it comes, feel more satisfying. The deceptive cadence is an example of the musical pleasure of having one's expectations played with skillfully — not frustrated (that would be unpleasant) but artfully redirected.

🔵 Try It Yourself: Finding Deceptive Cadences

Listen to the final chorus of Beethoven's Symphony No. 5, first movement, at approximately the 5-minute mark. Notice how several times the music seems to be ending — the dominant arrives, the trumpets blare, and... it's not over. The expected tonic resolution is repeatedly deferred through extended cadential progressions before the final resolution is allowed. This is the most famous example of extended cadential deferral in the Western canon. What is the emotional effect of this repeated non-arrival? How does the eventual resolution compare in emotional weight to what it would have been without the repeated deferrals?


27.7 Emotional Valence: What Makes Music "Sad" vs. "Happy"

The dimension of emotional valence — the positive/negative axis of emotion — is one of the most studied in music psychology. The acoustic correlates of valence are well-established, though the mechanisms by which they generate valence remain contested.

Major/Minor Mode

The most powerful single predictor of perceived valence in Western music is mode: major keys are generally perceived as positive, minor keys as negative. This association is so strong that even young children in Western cultures make the association reliably. Whether it is universal or culturally specific is the subject of Chapter 28, but within Western cultural contexts, it is extremely robust.

Tempo

Faster tempos are generally associated with higher valence (more positive) and higher arousal. Slower tempos are associated with lower valence and lower arousal. The acoustic basis of this association connects to the prosody of emotional speech: happy, energized speech is faster; sad, depressed speech is slower.

Register and Melodic Contour

Higher pitch registers tend to be associated with more positive valence; lower registers with more negative. Ascending melodic contours tend toward positive valence; descending toward negative. Again, these correlate with emotional speech prosody: enthusiastic speech rises in pitch; grieving speech falls.

Spectral Brightness and Timbre

Instruments and timbres with more high-frequency energy (spectral brightness) are associated with positive valence; darker timbres (more low-frequency energy, less high-frequency content) with negative valence. This is why a major chord played on a bright trumpet feels more cheerful than the same chord played on a muted, mellow horn.

📊 Data/Formula Box: Acoustic Correlates of Valence and Arousal

Acoustic Feature High Valence Low Valence High Arousal Low Arousal
Mode Major Minor
Tempo Fast Slow Fast Slow
Melody Ascending Descending Large leaps Small steps
Dynamics Loud Soft Loud Soft
Register High Low High Low
Spectral centroid Bright Dark Bright Dark
Articulation Staccato Legato Staccato Legato

27.8 Arousal: What Makes Music "Exciting" vs. "Calm"

The dimension of arousal — the activating/deactivating axis of emotion — is independently modulated by a different set of acoustic parameters. The two-dimensional circumplex model of emotion (Russell, 1980) proposes that most emotional states can be located on a two-dimensional space defined by valence and arousal, and musical emotion maps cleanly onto this space.

Tempo

Tempo is the single strongest predictor of perceived arousal. Fast music is arousing; slow music is calming. The acoustic basis is again prosodic: physiological arousal in humans is marked by faster speech rate; the emotional contagion mechanism (see Chapter 26) generates corresponding arousal in listeners.

RMS Energy (Loudness)

The root mean square energy of the audio signal — a measure of its average loudness — is a strong predictor of arousal. Loud music is more arousing than quiet music. This reflects both the evolved startle response to loud sounds (brainstem reflex, per BRECVEMA) and the social associations of loud music with high-energy, communal activities.

Spectral Centroid

The spectral centroid — the average frequency weighted by amplitude, a measure of the "brightness" of a sound — predicts both valence and arousal. Music with a high spectral centroid (lots of high-frequency energy) is experienced as both brighter/more positive and more arousing.

Rhythmic Complexity and Syncopation

Music with complex rhythms, heavy syncopation, and off-beat emphasis is more arousing than metrically simple music. This connects to the motor system: complex rhythms require more active temporal processing, engage motor prediction systems more intensely, and generate more of the embodied forward momentum associated with arousal.


27.9 The Spotify Dataset: Valence and Energy as Physical Measurements

🔗 In the Spotify Spectral Dataset (10,000 tracks, 12 genres), valence and energy are computed acoustic features assigned to every track by Spotify's audio analysis algorithms. These features operationalize the psychological dimensions described in sections 27.7 and 27.8, translating subjective emotional descriptors into objective acoustic measurements.

Spotify's Valence Feature

Spotify defines "valence" as a measure from 0.0 to 1.0 "describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry)." The algorithm is trained on audio features including mode (major/minor), tempo, spectral characteristics, and key — essentially, a weighted combination of the acoustic correlates described in Section 27.7.

Spotify's Energy Feature

Energy measures "a perceptual measure of intensity and activity," ranging from 0.0 to 1.0. It is computed from features including dynamic range, loudness (RMS), spectral entropy, and onset rate — aligning closely with the acoustic correlates of arousal described in Section 27.8.

What the Dataset Reveals

Analysis of the Spotify dataset across 12 genres reveals systematic genre-level patterns in the valence/energy space: - Electronic dance music (EDM): high energy, moderate-to-high valence - Classical music: low-to-moderate energy, widely distributed valence - Hip-hop: moderate-to-high energy, widely distributed valence - Ambient: low energy, widely distributed valence - Heavy metal: very high energy, low-to-moderate valence

These genre clusters reflect both the physical properties of the music (instrumentation, tempo, production style) and the cultural contexts in which the music is created and consumed. The alignment between acoustic features and emotional dimensions is strong enough that machine learning classifiers can predict genre from valence/energy alone with above-chance accuracy — though genre membership is also determined by many features that have no direct acoustic correlate (cultural history, lyrical content, identity).

⚠️ Common Misconception: Valence = Emotional Experience

Spotify's valence feature measures the emotional character signaled by the music, not the emotion experienced by the listener. A low-valence (sad-sounding) track might make one listener melancholic and another listener deeply pleased (because they enjoy sad music). The music-emotion link involves both the acoustic signal and the listener's state, history, and preferences. Valence features are useful for characterizing the acoustic space of emotion in music, but they should not be treated as predictions of emotional experience in individual listeners.


27.10 Embodied Emotion: Music as Movement Simulation

We touched on the embodied basis of rhythm in Chapter 26. Here we develop it more fully as a mechanism of musical emotion.

The evidence for embodied musical emotion is substantial:

Motor activation during listening: Neuroimaging studies consistently find motor cortex, premotor cortex, and supplementary motor area activation during passive music listening — regions involved in planning and executing movement, activated without any actual movement occurring.

Tempo and movement coupling: The tempo of music that people rate as most "pleasing" or "grooving" (approximately 120 BPM for many people) closely matches the tempo of comfortable walking — suggesting that musical pleasure is partly coupled to the ease with which the motor system can entrain to the rhythm.

Expressive gesture and emotion: The micro-timing variations that skilled performers introduce — the slight lengthening of melodic peaks, the subtle acceleration into cadences — directly parallel the kinematic profiles of human expressive movement. The musical performance "moves" in a way that the body recognizes.

The embodied account of musical emotion suggests that music's emotional power is partly the power of simulated motion. Music describes trajectories through a multidimensional space — pitch space, harmonic space, dynamic space — and the brain tracks these trajectories using the same systems it uses to plan and understand physical movement. The emotion generated is partly the emotion associated with those movement patterns in non-musical contexts: reaching upward (aspiration, tension), settling downward (resolution, completion), accelerating (excitement), decelerating (calm).

💡 Key Insight: The Kinesics of Music

"Kinesics" refers to the study of body movement as communication. Music, on the embodied account, is organized kinesics: it speaks in the language of movement to a brain that is built to interpret movement as emotionally significant. The reason a descending melody feels like "settling down" is that settling down — the physical act of lowering the body's center of mass — has the same felt quality. The reason an accelerating crescendo feels exciting is that acceleration and loudness are the kinematic and acoustic signatures of approach, which is exciting in virtually any context.

🔵 Try It Yourself: Conducting the Emotion

Without listening to any music, conduct imaginary music that expresses: (1) joy, (2) grief, (3) tension, (4) peace. Notice what your gestures naturally do in each case: how fast, how large, how smooth, how forceful. Then listen to a piece that exemplifies each of these states and compare your mental gestures to the actual musical features. The degree of alignment you find is evidence for the embodied account of musical emotion.


27.11 Social Emotion: Music and Shared Experience

Music is, across virtually all cultures, a social activity. It is made together, listened to together, danced to together, mourned and celebrated with. The question of whether musical emotion is inherently social — whether it requires, or is intensified by, the presence of other people — is answered clearly by the evidence: yes.

The Concert Context Effect

Studies comparing emotional responses to identical music in solo listening vs. concert contexts consistently find stronger emotional responses in concert settings. The presence of co-listeners who are visibly responding to the music — nodding, swaying, expressing emotion on their faces — provides social validation of one's own response and (via emotional contagion) amplifies the emotion itself.

Synchrony and Bonding

The neural synchrony described in Chapter 26 — the entrainment of listeners' neural oscillations to the music — has a social dimension: synchronized people tend to feel more bonded to each other. Experimental studies have found that people who are synchronized with each other (tapping in time together, rocking in chairs at the same rate) show increased cooperation, prosocial behavior, and feelings of social connection. Music, by synchronizing the bodies and nervous systems of a group, may be a powerful tool for social bonding — one of its most important evolutionary and cultural functions.

Communal Music-Making

The emotional experience of making music together — in a choir, an orchestra, a jam session, or a drum circle — appears to be qualitatively different from listening, involving higher oxytocin levels, stronger prosocial feelings, and a specific quality of "merging" with the group that solo performance does not produce. This may be why communal singing is found in virtually every known human culture and why it is used in contexts of social bonding — religious ritual, military marching, political protest, funeral rites.


27.12 Negative Emotions and Music: Why Do We Like Sad Music?

One of the most persistent puzzles in the psychology of musical emotion is the paradox of pleasurable sadness: people regularly report enjoying sad music, seeking it out when they are sad, and finding it comforting. How can sadness — which is, by definition, negatively valenced — be pleasurable?

The Prolactin Hypothesis

One proposed explanation involves the hormone prolactin, which is released in response to grief and is thought to counteract the negative feelings associated with loss. David Huron proposed that sad music triggers a neural response that is actually a grief response — the music "tricks" the brain into processing it as a real loss — and the resulting prolactin release produces a calming, comforting counteraction to the induced sadness. On this account, the pleasure of sad music is the pleasure of comfort in response to grief.

The Safe Context Hypothesis

Another explanation emphasizes the safety of the musical context: music provides a way to experience negative emotions without real consequences. The sadness induced by music is known to be temporary, controlled, and not associated with actual loss. This "aesthetic distance" allows the listener to explore and process negative emotional states without threat, which may itself be valuable — a form of emotional rehearsal.

The Social Surrogacy Hypothesis

A third account proposes that sad music functions as a social surrogate: it provides a sense of being accompanied and understood that is particularly valuable when one is actually sad and isolated. The composer's emotional expression, mediated through the music, feels like a form of companionship.

Aesthetic Appreciation

Finally, some listeners who enjoy sad music report that their response is not primarily sadness itself but aesthetic appreciation of the craft involved in its expression — admiration for how beautifully the music captures and expresses the quality of sadness. This positions their response in the "aesthetic emotion" category: not sadness but meta-sadness, the beautiful-sad feeling that is triggered by artistic excellence in the expression of grief.

These accounts are not mutually exclusive. The pleasure of sad music likely arises from a combination of several mechanisms, varying across listeners and listening contexts.


27.13 Thought Experiment: Design a Maximally Emotionally Effective Excerpt

🧪 Thought Experiment: The Optimally Moving Music

Using only the principles developed in this chapter, design a short musical excerpt (30–60 seconds) that would be predicted to generate the strongest possible emotional response in a typical Western listener. Specify:

  1. Key and mode: What key and mode? Does it modulate? If so, to where?
  2. Tempo and tempo variation: What is the tempo? Does it accelerate, decelerate, or both?
  3. Harmonic rhythm: How fast do the chords change? What is the progression?
  4. Melodic design: What does the melody do? Where are the peaks? How do they resolve?
  5. Dynamics: What is the dynamic arc? Where is the climax?
  6. Timbre and orchestration: What instruments? What register?
  7. ITPRA design: Where are the strongest expectation-building passages? Where is the biggest prediction violation? Where is the most satisfying resolution?

Now add a further constraint: the excerpt must also be aesthetically interesting — not merely mechanically manipulative. Can you satisfy both criteria simultaneously, or is there a tension between psychological effectiveness and aesthetic quality?

This thought experiment mirrors the actual compositional challenge. Great composers — from Bach to Adele — are both technically skilled in manipulating the acoustic correlates of emotion and artistically skilled in deploying those techniques in ways that feel inevitable rather than calculated.


27.13b The Spotify Dataset Deep Dive: Genre Clusters in Emotional Space

🔗 Returning to the Spotify Spectral Dataset (10,000 tracks, 12 genres), valence and energy measurements across the full dataset reveal patterns that illuminate how musical emotion is constructed and consumed at scale.

Genre-Level Emotional Signatures

When the 10,000-track dataset is visualized as a scatter plot with valence on the x-axis and energy on the y-axis, distinct genre clusters emerge:

High energy, moderate-to-high valence (upper right quadrant): EDM, pop-dance, and hip-hop tracks with bright, fast, loud production cluster here. These are the most commonly used tracks in exercise contexts and social settings. Their acoustic features maximize the brainstem reflex and rhythmic entrainment BRECVEMA mechanisms.

High energy, lower valence (upper left quadrant): Heavy metal, aggressive hip-hop, and hardcore punk tracks cluster here. High arousal combined with negative valence is the signature of anger-related music — fast, loud, but minor-key or modal, with lyrical themes of aggression or protest. Listeners of this music report energizing effects, consistent with the paradox that high arousal can feel pleasant even when valence is negative.

Low energy, moderate-to-low valence (lower left quadrant): Singer-songwriter, acoustic indie, and some classical music clusters here. This is the quadrant of calm contemplation — quiet, introspective music with mixed or negative valence that invites the Default Mode Network's inward-oriented processing.

Low energy, higher valence (lower right quadrant): Acoustic folk, gentle country, and some ambient music. This is the quadrant of contented calm — pleasant but not arousing. The least commercially dominant quadrant for streaming but important for background and focus use cases.

What the Genre Clusters Tell Us About Emotion Production

The systematic relationship between genre and emotional space is not accidental. Genres are socially organized communities of practice that have, over decades or centuries, evolved toward specific emotional functions. EDM maximizes arousal and positive valence because it evolved in dance contexts where both are functional. Heavy metal maximizes arousal while tolerating or embracing negative valence because it evolved in contexts where cathartic anger-expression was the function. Singer-songwriter music cultivates low-energy introspection because it evolved in contexts of intimate self-expression and listening.

The Spotify acoustic features provide a way to see this cultural evolution reflected in physical measurements — a bridge between cultural history and acoustic physics.

The Valence Measurement Problem

One significant limitation of Spotify's valence feature deserves attention: it is calibrated primarily on Western popular music and may not accurately reflect the emotional valence of music from other traditions. A high-valence track in Western pop is one that sounds "happy" by the acoustic standards of that tradition — major key, fast, bright. A high-valence track in a raga tradition might have completely different acoustic features. The Spotify valence feature is an acoustic measurement tool calibrated to a specific cultural musical vocabulary, not a universal emotional measurement instrument.

This limitation is an instance of the broader universal-vs-cultural theme: the acoustic features that signal positive valence are themselves culturally specific, even when measured as objective acoustic properties. An instrument calibrated on one cultural musical tradition will produce systematic errors when applied to another.

27.14 The Anatomy of a Musical Climax: Case Analysis

To make the abstract principles of this chapter concrete, consider the approach to and achievement of the climax in a large-scale musical work — specifically, the final approach to the principal climax in the first movement of Beethoven's Fifth Symphony.

The Setup: Measures 1–390 (approximately)

The entire first movement is a study in accumulated tension. The famous four-note motif (short-short-short-long, G-G-G-E♭ in C minor) is stated in the opening measures and then relentlessly developed, fragmented, inverted, and intensified throughout the movement. The harmonic language sustains tension through long passages in the minor mode, repeated half-cadences (ending on the dominant, refusing resolution), and sudden dynamic contrasts (the explosive ff following whispered pp passages).

The Pre-Climax Buildup: Measures 390–440 (approximately)

As the recapitulation approaches its final section, the harmonic rhythm slows, the dynamics build, and the bass moves inexorably toward a dominant pedal — a long sustained note on G (the dominant of C minor) over which the entire orchestra builds a climactic crescendo. The dominant pedal is one of the most powerful tension-building devices in Western music: by sustaining the harmonically tense dominant note for an extended period, it creates an irresistible pressure toward tonic resolution that cannot be denied.

During this buildup, every neural prediction system is active: - The harmonic expectation system predicts C major/minor arrival (tonic) - The melodic expectation system anticipates the main theme's return at the tonic - The dynamic expectation system, trained on the entire movement, expects a massive forte arrival - The temporal prediction system anticipates arrival on a strong metric downbeat

The Climax: The Moment of Resolution

When the recapitulation's tonic arrival occurs — when the C minor triad arrives in the full orchestra on the downbeat after all the dominant buildup — all of these prediction systems confirm simultaneously. The reaction phase of the ITPRA cycle fires at maximum intensity for all systems at once. This convergent confirmation is the physical basis of the overwhelming subjective sense of arrival, inevitability, and satisfaction that makes this moment one of the most recognizable in Western music.

But Beethoven adds a further layer: immediately after the tonic arrival, he introduces a new element (a short oboe solo in the Classical style, completely unlike the rest of the movement's character) that creates a new, unexpected prediction violation within the moment of resolution. This violation — the surprise of beauty within the moment of resolution — triggers a fresh dopamine response and creates the sense of the climax being both conclusive and somehow beyond expectation.

📊 Data/Formula Box: Forces at Work at a Musical Climax

Force Physical Mechanism Emotional Effect ITPRA Phase
Dominant pedal Sustained dissonant bass note Accumulated tension Imagination → Tension
Dynamic crescendo Increasing RMS energy Arousal intensification Tension
Rhythmic unison All voices synchronized on the beat Maximal motor entrainment Prediction
Tonic arrival Dissonance → consonance; expectation confirmed Release + reward Reaction
Surprise element New prediction violation within resolution Second-order pleasure Appraisal

27.14b The Acoustic Correlates of Musical Tension: A Formal Summary

Having examined harmonic, melodic, and rhythmic tension-release separately, it is useful to gather the acoustic correlates into a unified picture. Musical tension, as a subjective psychological state, is predicted by the following measurable acoustic quantities:

📊 Data/Formula Box: Acoustic Predictors of Musical Tension

Acoustic Measure Mathematical Quantity Direction Contribution
Spectral roughness Sum of beating between all harmonic pairs Increases → tension Primary (harmonic tension)
Melodic height (mean pitch) Mean fundamental frequency of melody Increases → tension Secondary
Melodic interval size Mean absolute interval between successive notes Increases → tension Secondary
Distance from tonic in pitch class space Krumhansl tonal hierarchy distance Increases → tension Primary (tonal)
Metrical strength of current beat Position in metrical hierarchy Decreases → tension Primary (rhythmic)
RMS energy √(mean squared amplitude) Increases → tension/arousal Mixed
Onset density Number of note onsets per second Increases → tension Secondary
Harmonic rhythm Rate of chord change Increases → tension Secondary

The key insight is that tension is not a single quantity but the weighted sum of multiple partially independent physical quantities, each of which contributes to the overall subjective tension experience. A passage can be harmonically tense but rhythmically resolved; melodically tense but dynamically quiet. Great composers control these dimensions semi-independently, creating complex textured tension profiles that guide the listener through multi-dimensional emotional space.

Fred Lerdahl's Tonal Pitch Space (2001) provides the most rigorous formal treatment of tonal tension, computing a numerical "tonal tension" value for every note event in a tonal piece based on its distance from the local and global tonic in a hierarchically structured pitch-class space. Studies have found correlations of approximately 0.7–0.8 between Lerdahl's computed tension values and listener tension ratings — a strong result that demonstrates the feasibility of a quantitative account of musical tension, while the imperfect correlation (0.3–0.2 unexplained variance) confirms that acoustic-physical factors do not fully capture subjective tension.

27.14c Individual Differences in Musical Emotional Response

The acoustic correlates of emotion predict average group responses — they are statistical regularities. Individual listeners vary enormously in their emotional responses to the same music, for reasons that are partly biological, partly developmental, and partly situational.

Empathy and Emotional Contagion

Tillie Eerola and colleagues found that measures of empathy — specifically "affective empathy" (the tendency to directly share others' emotional states) — predict the strength of emotional response to sad music. High-empathy individuals show stronger physiological and subjective responses to emotionally expressive music, consistent with the emotional contagion mechanism being stronger in people who are generally more sensitive to emotional signals from others. This suggests that the emotional contagion mechanism (BRECVEMA's "C") varies substantially across individuals based on underlying social-affective sensitivity.

Absorption and Musical Engagement

The psychological trait of "absorption" — the tendency to become deeply involved in engaging experiences — predicts the intensity of musical emotional response across genres. High-absorption individuals are better at achieving the kind of sustained, focused attention that allows musical expectation to build and resolve at maximum intensity. Low-absorption individuals may "pull out" of musical engagement mentally, missing the built-up expectation that makes climactic moments maximally effective.

Current Emotional State (Mood Priming)

A listener's emotional state before hearing music significantly influences their emotional response to it. Negative mood tends to increase the emotional response to sad music (the social surrogacy mechanism is more salient when one is already sad) but can decrease the response to energizing, positive music (the gap between the music's positive signal and the listener's negative state creates a kind of cognitive dissonance). Positive mood tends to amplify responses to happy music and may buffer against the full impact of sad music.

Musical Expertise and Expectation Richness

Musical training produces richer and more precise musical expectations. A trained musician hearing a modulation to a distant key has a more precisely specified expectation system that is more dramatically violated — the prediction error is larger relative to the prior. This may explain why professional musicians often report stronger emotional responses to sophisticated, harmonically complex music than non-musicians, even though non-musicians may respond more strongly to simple, metrically regular, major-key music that aligns with their simpler expectations.

27.14d Cultural Variation in Musical Emotional Response: The Universal-Cultural Interaction

The cross-cultural dimension of musical emotional response is complex: some features are relatively universal; others are strongly culturally specific.

More Universal Acoustic Correlates

Tempo is the most cross-culturally reliable correlate of arousal: fast music is experienced as more energizing than slow music in virtually every studied musical culture. This likely reflects both the evolved association between acoustic velocity and physical movement speed and the universal coupling of rhythmic auditory stimulation to the motor system.

Dynamic intensity (loudness) is similarly cross-culturally reliable as an arousal cue: louder music is more arousing. This reflects both the acoustic startle component of loud sounds and the cultural universality of the association between intensity of sound and intensity of event.

More Culturally Specific Correlates

Mode (major/minor) is the acoustic feature most strongly subject to cultural specification, as Chapter 28 will document at length. The correlation between minor mode and negative valence is very strong in Western populations and substantially weaker in many non-Western populations.

Specific tonal and melodic conventions — the dominant-tonic relationship, the leading tone's directional drive, the emotional character of specific chord progressions — are largely specific to Western tonal tradition. A cadential 6/4 (the tonic chord in second inversion that precedes the dominant in classical cadences) creates intense tension for a Western-trained listener; it is acoustically unremarkable to a listener without Western tonal training.

The Interaction: Physical Seed, Cultural Amplification

The universal-to-cultural gradient in acoustic emotional correlates mirrors the account developed throughout Part VI: acoustic properties provide cross-cultural seeds (tempo-arousal, loudness-arousal, roughness-roughness) that are then amplified, refined, and extended by cultural learning into the rich emotional landscape that any specific musical tradition offers its listeners. Western listeners, after centuries of cultural development and individual enculturation, have a uniquely dense and specific set of music-emotion associations that go far beyond what acoustic properties alone could generate — but that also cannot be understood without reference to those acoustic properties.

27.15 Theme 1 Checkpoint: Is Musical Emotion Reducible to Acoustic Physics + Prediction?

⚖️ Debate/Discussion: Is musical emotion "real" emotion or a specialized aesthetic response?

The account of musical emotion developed in this chapter is impressive in its scope and coherence. We have acoustic features that predict emotional valence and arousal. We have the BRECVEMA model identifying eight distinct mechanisms. We have the predictive coding framework (ITPRA) explaining the temporal dynamics of emotional response. We have the embodied account connecting music to movement and the social account connecting it to bonding.

Is this enough? Is musical emotion — the moved-to-tears response of the Carnegie Hall audience during Górecki's symphony — fully explained by this account?

The Case That It Is

A committed physicalist argues yes: musical emotion is the integrated output of the mechanisms described in this chapter, operating on a brain shaped by evolution and culture. The Górecki audience's tears are the phenomenal expression of elevated prolactin and opioids, of prediction errors resolving in the slow, inevitable arc of a symphony, of embodied simulation of the music's expressive gestures, of evaluative conditioning connecting slow string music to grief, of episodic memories triggered by the music's stylistic associations. There is no "more" to explain.

The Case That There Is Something Irreducible

The philosophical rejoinder is familiar: even a complete causal-mechanistic account of why the Górecki audience cried does not capture what it is like to be moved by that symphony. The felt quality of grief-and-beauty that the symphony produces — the specific phenomenal character of that response — is not captured by any physical description of the acoustic signal or the neural response. This is not anti-scientific; it is the hard problem of consciousness applied to music, and it remains genuinely open.

A Practical Resolution

For the practicing musician, neuroscientist, or educator, the account in this chapter provides powerful and actionable knowledge: understanding the acoustic correlates of emotion allows compositional and analytical insight; understanding the mechanisms allows therapeutic and educational applications; understanding the limits of prediction allows appropriate humility about the irreducible particularity of individual musical experience.

The distinction between "fully explained" and "partially understood" may matter less than the question of what the partial understanding enables. In that sense, the physics and neuroscience of musical emotion — however incomplete as a final account — is enormously valuable.


27.15b Neuroimaging Evidence for Tension and Release

The subjective experience of musical tension and release has correlates in brain activity that can be measured with neuroimaging. Studies combining fMRI with continuous tension ratings from listeners have found that:

Auditory Cortex

Primary and secondary auditory cortex show increasing BOLD signal as harmonic tension increases. This is consistent with the roughness account: more spectral roughness → more complex auditory processing → greater neural activity in auditory cortex.

Nucleus Accumbens and Caudate

As discussed in Chapter 26, these reward-related regions show increased activity during the buildup before musical climaxes (anticipatory tension) and a further increase at the moment of resolution (reward receipt). The temporal profile of nucleus accumbens activation tracks the ITPRA cycle: it rises during the Imagination and Tension phases, peaks during the Reaction phase (at resolution), and decays during Appraisal.

Supplementary Motor Area and Premotor Cortex

Motor planning regions show increased activity during rhythmically complex, syncopated passages — the passages that score highest on the "rhythmic tension" dimension. This is consistent with the embodied account: rhythmically tense music engages the motor system in a state of heightened preparatory activation (motor prediction without motor execution).

Anterior Cingulate Cortex

The anterior cingulate cortex (ACC) — a region involved in conflict monitoring, error detection, and emotional arousal — shows elevated activity during harmonically tense passages and at moments of harmonic violation (deceptive cadences, unexpected modulations). This ACC activity may reflect the neural signature of prediction error — the "conflict" between the predicted and actual harmonic event.

Prefrontal Cortex

Lateral prefrontal cortex shows increased activity during harmonically complex passages, consistent with its role in working memory and executive processing — holding the current harmonic context in mind while evaluating new harmonic events. The extent of prefrontal engagement correlates with musical training: more sophisticated listeners show more prefrontal engagement during complex passages, consistent with richer contextual processing.

27.15c Tension-Release in Non-Western Musical Traditions

Tension-and-release is not uniquely Western. All known musical traditions exploit the psychological dynamic of tension and resolution — but the physical means by which they create and resolve tension vary considerably.

Rhythmic Tension in Indian Classical Music

Indian classical music (Hindustani and Carnatic traditions) creates extraordinary tension through rhythmic means that have no exact Western equivalent. The tala system — a cycle of beats with specific metric weight patterns — sets up a rhythmic expectation cycle. During improvisation, a performer might play patterns that displace notes across multiple tala cycles, creating complex cross-rhythms whose resolution back to the "sam" (the first beat of the tala cycle) can be felt as an overwhelming rhythmic release. This rhythmic tension-resolution is created entirely without harmonic manipulation; it is purely temporal.

Ornamentation and Tension in Bluegrass and Country

In bluegrass and traditional country music, melodic tension is created through the "blue note" — a flattened third, fifth, or seventh that clashes with the major-key context. This acoustic roughness creates a brief moment of tension that is immediately resolved when the melody moves to the adjacent diatonic note. The emotional effect — a bittersweet, slightly aching quality — is the characteristic emotional color of these traditions, created through a minor-key inflection in a major-key context. This is tension-and-release at the scale of a single ornamental note.

Tension and Silence in Japanese Ma

Japanese traditional music (including shakuhachi, koto, and traditional theater music) exploits silence as a tension device in ways that Western music rarely does. The concept of ma — the meaningful pause, the pregnant silence — creates a form of temporal tension that is resolved only by the next sound. The silence is not the absence of music but an active musical event, charged with anticipation. This is a form of ITPRA tension without acoustic content: the Imagination and Tension phases are sustained through silence, making the eventual Reaction even more powerful.

These cross-cultural examples confirm that the basic psychological mechanism of musical tension-release — expectations created, maintained, and resolved — is universal, while the physical means of creating and resolving tension are culturally various. This is consistent with the chapter's overall framework: universal psychological mechanisms, culturally specific acoustic realizations.

27.15d The Philosophical Questions at the Heart of Musical Emotion Research

As we approach the end of this chapter, it is worth confronting directly the deepest philosophical difficulties in the science of musical emotion.

The Reduction Problem

The most comprehensive account of musical emotion that science currently offers — the BRECVEMA model plus the predictive coding framework plus the neural correlates — is impressive in its scope. It identifies eight distinct causal mechanisms, traces their neural implementation, and relates them to the acoustic features of music. Can we conclude that this account explains musical emotion?

A reductionist would say yes: musical emotion just is the integrated output of these mechanisms, operating in a brain and body of a particular history. There is nothing more to explain.

A philosopher sympathetic to the "intentionality" of emotions would note that the reductionist account leaves out the intentional object of the emotion: what the emotion is about. When you feel moved by a musical passage, you are moved by the music — by its specific qualities, its combination of elements, its arrival at a particular moment. The neural account describes the mechanism; it does not capture the intentional structure of the emotional experience.

The Problem of Expression

Music is often described as "expressing" emotions: a piece expresses sadness, or hope, or triumph. But what does it mean for a physical object (organized sound) to express an emotion? Three answers are on offer:

Arousalist: Music expresses an emotion because it causes that emotion in listeners. But this conflates expressing with causing.

Resemblance: Music expresses an emotion because it resembles the acoustic or kinematic features of human emotional expression. This is the prosody hypothesis and the embodied account. But it doesn't explain why a piece with no vocal-sounding timbres (e.g., a purely abstract electronic composition) can express specific emotions.

Convention: Music expresses emotions through learned conventions — minor mode expresses sadness because we've been taught that it does. But this makes expression arbitrary in a way that doesn't capture the felt inevitability of musical expression.

None of these accounts is fully satisfying, and the most sophisticated current accounts acknowledge that music expression is likely a compound of all three. The scientific accounts in this chapter primarily illuminate the arousalist mechanism and the resemblance mechanism; the convention mechanism is addressed in Chapter 28.

What Science Has Accomplished

Despite these philosophical open questions, the science of musical emotion has made genuine progress: it has identified specific causal mechanisms with specific neural implementations; it has mapped the acoustic correlates of emotional dimensions with enough precision to be practically useful (in music therapy, in film scoring, in music recommendation systems); it has revealed cross-cultural patterns that constrain theories of musical universals; and it has provided clinical insights that have real benefits for patients with neurological and psychiatric conditions. This is substantial knowledge, even if it is not — and perhaps cannot be — the complete story.

27.16 Summary and Bridge to Chapter 28

Musical emotion is not a single thing but a family of related processes, each with distinct mechanisms, timescales, and relations to acoustic features. The BRECVEMA model organizes eight of these mechanisms into a coherent taxonomy; the ITPRA framework explains the moment-to-moment temporal dynamics of musical emotional response; the acoustic correlates of valence and arousal provide a physical basis for the emotional dimensions that cross-cultural music psychology consistently identifies.

At the center of all these accounts is the concept of expectation: musical tension is predictive tension, musical release is predictive confirmation, and the pleasure of music is largely (though not entirely) the pleasure of a sophisticated, real-time prediction system being engaged, challenged, and ultimately satisfied. The brain that evolution built for navigating a complex, temporally structured world turns out to be ideally equipped to find pleasure in the complex, temporally structured phenomenon of music.

Key Takeaways

  • Musical emotion is not a single phenomenon but a family of related responses produced by multiple mechanisms (per the BRECVEMA model).
  • Huron's ITPRA framework explains the temporal dynamics of musical emotion as a sequence of imagination, tension, prediction, reaction, and appraisal — a cycle that repeats at every temporal scale.
  • The acoustic correlates of emotional valence include mode, melodic direction, tempo, and spectral brightness; those of arousal include tempo, RMS energy, and spectral centroid.
  • Musical tension has physical correlates: harmonic dissonance (beating frequencies), melodic height and direction, and rhythmic displacement from the metric pulse.
  • The deceptive cadence is the paradigm case of emotion-through-violated-expectation: expectation is fully built, then artfully redirected rather than fulfilled.
  • Embodied accounts of musical emotion connect acoustic motion through pitch/dynamic space to the physical movement patterns associated with emotional states.
  • The social dimension of musical emotion — the amplifying role of communal listening and music-making — is underappreciated in individualistic laboratory accounts of musical response.
  • The paradox of sad music is not a paradox but a multiplicity: pleasurable sadness arises from combinations of prolactin-mediated comfort, safe emotional exploration, social surrogacy, and aesthetic appreciation.

Bridge to Chapter 28

Among all the acoustic correlates of emotional valence, none is more famous, more debated, or more culturally complex than the major/minor distinction. Chapter 28 takes this single phenomenon — minor sounds sad, major sounds happy — and submits it to the full range of physical, cultural, cognitive, and developmental analysis. The answer turns out to be more complicated than either "it's just physics" or "it's just culture" — and the complications are where the most interesting science lives.