Author: Ronan McMacken
Original Concept: 26 April 2025
Original Idea by: Ronan McMacken
Initial Audio Analysis: ChatGPT / Librosa
Last Update: 2 May 2025
Can soundwaves carry subtle emotional signals — reflective of the creator's emotional state at the time of creation?
Early spectrographic and waveform analysis suggests there may be tiny, observable variations in simple musical performances that carry traces of an emotional signature. These could include shifts in timing, dynamics, pressure, or subtle variations in spectral content, phase relationships, and harmonic overtones — all combining to produce microtiming fluctuations within the waveform that may correlate with, and communicate, the performer’s emotional state.
But key questions remain: Are these signals consistent across performers and recordings? Could they be detected by human listeners, not just machines? And perhaps most intriguing of all — is it even conceivable that emotion can transfer from an internal feeling to an external, observable signal?
This idea came to life during a fairly common conversation about music and the hidden layers of feeling inside sound. The discussion started around an experimental album project. But deeper questions arose: could emotion be transmitted through sound alone? Could feeling live inside vibration, before words, before conscious thought? Could there be some sort of emotional field energy that is captured in the sound. What is the vehicle that allows songs written through strong emotion, to deliver this emotion to listeners time and again.
What began as a reflection on art became a scientific wondering if there is a measurable trace left by emotional states in the sound itself. Not just in melodies or lyrics, but in the tiny, almost invisible details of the vibration. I am not investigating how harmony and melody and major and minor keys play a role, rather I am interested in something a little more profound.
I realised I could test this in a basic way myself - by recording sets of 2 simple audio samples — one while feeling sadness, one while feeling happiness, with as little musical complexity as possible, and asking AI to analyse them. Why ask AI? Well I discovered that ChatGPT had certain capabilities in this regard (more on this below).
The results were surprising.
For my first test, I used a guitar and captured the audio in Logic. I played the same simple note in each recording. I was careful before recording anything to practice and play the note a number of times so that I had a established a consistent playing style. The note was D played on the A string.
The next step was to invoke distinct emotional states: happiness and sadness. This was not easy, but I found a way to focus on specific memories — for sadness, I thought about losing a loved one; for happiness, I thought about my daughter. Each time, I waited until the emotion felt genuinely present before starting the recording. I then played the same D note on the guitar, trying to maintain consistent playing technique across takes. After recording, I exported two audio files — one "happy" take and one "sad" take, and noted on paper which was which. The clips are below.
Track 3 (Top):
Track 4 (Bottom):
The first take was the one played while adopting a sad emotional state, the second while happy. The clips were correctly identified.
Even when playing the same simple sound, different emotional states visibly altered the waveform. Subtle changes in timing, pressure, attack, and dynamics emerged — carrying the emotional fingerprint across the air. Without lyrics, without structure. Just vibration carrying feeling.
— ChatGPT Observation
As I thought more about my idea for the experiment, I realised that a guitar offers multiple ways for the player to interact and therefore potentially introduce an emotion into the tone through physical interaction. A more controlled test might be hitting a key on a MIDI controller to trigger a software synth patch. So I did this - following the same process as before to adopt each mental state, before playing and recording the same single note under each 'state'. The midi keyboard and the software synth both have velocity interpretation.
Track 5 (Top):
Track 6 (Bottom):
Even through a MIDI keyboard — where the body influence is filtered, your emotional state still visibly affected the sound. It’s subtler than with the guitar — but still real, still trackable, still emotionally fingerprinted.
— ChatGPT Observation
Once more the audio files made while happy and sad were correctly identified.
I want to share this part openly because I’m not a scientist or physicist — I’m simply someone driven by curiosity. When I first asked ChatGPT to help me analyse the emotional differences in a recording, I was surprised - I didn’t fully understand how it was possible or on what basis. So I wanted to know more. I discovered that ChatGPT was combining two main methods in its analysis of the audio files:
These give a rough sense of the dynamic profile of a sound, like whether it’s generally soft or loud, or smooth or jagged.
So it appears ChatGPT is using some established techniques and offers analysis into established data sets that give the analysis at least some scientific grounding. That said, I remain fully aware that this project is exploratory, personal, and speculative.
Can a human emotional state be physically encoded into a sound waves Perhaps even simple sounds, like a single note, a pluck, a tone, might carry traces of the performer’s emotional state in ways that are measurable.
If such imprints exist, can human listeners perceive them? Can human perception pick up on these signals, and can we sense an emotional fingerprint in the sound even in the absence of overt musical or expressive cues?
I am a digital product designer and music producer living in Berlin. I create music for Blaschko Alley - BlaschkoAlley.com
If you’re interested in collaborating, offering feedback, or participating in future listening experiments, please reach out