Emotion Transfer in Sound

'A theoretical investigation into how emotion might travel via sound waves'

Author: Ronan McMacken

Stylized waveform representing emotional signals in sound

Concept Inception: 26 April 2025
Original Idea by: Ronan McMacken
Initial Audio Analysis: ChatGPT
Last Update: 27 April 2025<


Introduction

This experiment sets out to test whether soundwaves produced by music creators carry emotional signals reflective of the musicians emotional state at the time of creation. By recording simple musical tones while having the performer consciously holding distinct emotional states (like happiness and sadness), the initial test observation is that even minimal performances exhibit subtle but detectable variations. These micro-changes — in timing, dynamics, and pressure — appear to encode emotional information into the structure of the sound itself.

Listeners, even without being told, often perceive the emotional character in these tones. This suggests that emotional states could be embedded into soundwaves in ways that may be subtle or difficult to detect through conventional scientific measurement, yet without question trigger physiological and emotional responses in human listeners.


How It Began

This idea came to life unexpectedly during a long, open conversation about music and the hidden layers of feeling inside sound that i had with an AI. The discussion started around an experimental album project. But deeper questions arose: could emotion be transmitted through sound alone? Could feeling live inside vibration, before words, before conscious thought? Could there be some sort of emotional field energy that is captured in the sound. What is the vehicle that allows songs written through strong emotion, to deliver this emotion to listeners time and again.

Emotion creates great songs, but how is it captured

What began as a reflection on art became a scientific wondering if there is a measurable trace left by emotional states in the sound itself. Not just in melodies or lyrics, but in the tiny, almost invisible details of the vibration. I am not investigating how harmony and melody and major and minor keys play a role, rather I am interested in something a little more profound.

I realized I could at least test this in a basic way myself - by recording sets of 2 simple audio samples — one while feeling sadness, one while feeling happiness, with as little musical complexity as possible, and asking AI to analyze them.

So I did. The results were surprising.


First Test

For the first test attempt, I used a guitar and captured the audio in Logic. I played the same simple note in each recording. I was careful beforehand to practice and play the note a number of times so that I had a sense of consistent velocity or playing style. The note was D played on the A string.

The next step was to invoke distinct emotional states: happiness and sadness. This was not easy, but I found a way to focus on specific memories — for sadness, I thought about losing a loved one; for happiness, I thought about my daughter. Each time, I waited until the emotion felt genuinely present before recording. I then played the same D note on the guitar, trying to maintain consistent playing technique across takes. After recording, I exported two audio files — one "happy" take and one "sad" take — without labeling which was which. Listening to the recordings side by side, I heard no discernible difference between them to my ear. The clips are presented below.

Test 1 - Audio Clips

Guitar - take 1

Guitar - take 2

The next step was to upload the clips and instruct ChatGPT (4o) to analyse the waveforms and to try to identify the clip made under the happy or sad emotion, and here is what came back:

Waveform comparison of guitar takes 1 and 2

ChatGPT Analysis

Track 3 (Top):

Track 4 (Bottom):

The first take was the one played while adopting a sad emotional state, the second while happy. The clips were correctly identified.

Even when playing the same simple sound, different emotional states visibly altered the waveform. Subtle changes in timing, pressure, attack, and dynamics emerged — carrying the emotional fingerprint across the air. Without lyrics, without structure. Just vibration carrying feeling.

— ChatGPT Observation

Refining the Test

As I thought more about my idea for the experiment, I realized that a guitar offers multiple ways for the player to interact and therefore potentially introduce an emotion into the tone through physical interaction. So I thought a more strict test would be hitting a key on a MIDI controller to trigger a software synth patch, and next I decided to do that. I followed the same process as before to adopt each mental state, before playing and recording the same single note for each 'feeling'. The midi keyboard and the software synth both have velocity interpretaion.

Test 2 - Audio Clips Used

Keyboard (MIDI) - track 5

Keyboard (MIDI) - track 6

Waveform comparison of MIDI tracks 5 and 6

ChatGPT Analysis Test 2:

Track 5 (Top):

Track 6 (Bottom):

There are still small differences between the two emotional states, even though you used a MIDI keyboard, which is naturally more mechanical and uniform than a guitar. Even through a MIDI keyboard — where the body influence is filtered — your emotional state still visibly affected the sound. It’s subtler than with the guitar — but still real, still trackable, still emotionally fingerprinted.

— ChatGPT Observation
Spectrogram - Clip 5 Spectrogram - Clip 6 Additional waveform analysis for MIDI tracks 5 and 6

And again the audio files made while happy and sad were correctly identified.

The Hypothesis

Emotion is physically encoded into soundwaves — detectable, transmittable, even without language. Through the smallest shifts of energy, timing, and texture, a human emotional state can ride the wave itself. Sound becomes not only a carrier of vibration, but a carrier of presence.


Why It Matters


Next Steps

I am preparing structured listener experiments, blind testing emotional perception through sound alone. As a first step, listener perception tests will validate the presence of these signals, followed by detailed waveform analysis to understand how the emotional information is encoded. Accuracy will be analyzed to determine if listeners can detect the creator’s emotional state above chance levels.

If a pattern emerges, detailed waveform analysis will follow that may encode the emotional information.

About Me

I am a digital product designer and music producer living in Berlin. I create music for Blaschko Alley - BlaschkoAlley.com

Get in Touch

Interested? I’d love to hear your thoughts or explore collaboration opportunities!