Hacker News new | past | comments | ask | show | jobs | submit login
Everything we can't describe in music (hazlitt.net)
139 points by anarbadalov 6 months ago | hide | past | favorite | 108 comments



> "Timbre’s role in music has always been underrated, or even ignored, probably because it is an intangible that’s difficult to describe, hard to categorize, and so far, immune to measurement. "

In the mid-to-late 1800s, Herman Helmholtz was building spectrographs of musical sounds by listening to them through a bunch of resonators of varying sizes that acted as band-pass filters, and recording how loud each harmonic sounded.

His writings are still relevant today.

https://www.amazon.com/Sensations-Tone-Dover-Books-Music/dp/...

I've heard that the Sethares book (which is much more recent) is really good too, but I don't have a copy of that one.

https://www.amazon.com/Tuning-Timbre-Spectrum-William-Sethar...


The quoted section of the article is pretty out of touch with modern music. Many (most?) genres of EDM place the importance of timbre and texture above harmony and melody


That was mentioned in the article.

> Timbre is even more crucial in contemporary genres such as electronic dance music and rap. EDM, said Swoger-Ruston, is really “experimentation with timbre.” The widespread use of sampling means the sample, not the note, is increasingly the basic unit of music.


I think where the article is out-of-touch is implying that timbre is still largely a mystery, when actually, for the most part, it isn't.


Yeah, well frankly most EDM fucking sucks and that's pretty much why. Your average cool guy "producer" doesn't know the first thing about intervals, counterpoint, or even basics like triads or major/minor scales. But they sure can twist knobs in a way that makes neat sounds!


First of all, who cares?

But that aside, this take reeks of sampling bias. Every amateur with a laptop uploads their music to SoundCloud/Youtube/whatever even while they are still learning. You probably don’t have much exposure to the crappy piano fugues and poorly orchestrated droning sonatas that are written by early university students. Trust me there’s a lot of that as well. Everyone starts somewhere after all.


Judging a piece based on its technical merit or the artist’s education is extremely limited and pretentious. There’s a tremendous amount of beautiful music made on laptops by people who have no formal training or knowledge of theory. Not everything full of synthesisers and samples is “EDM”.


> Not everything full of synthesisers and samples is “EDM”.

I didn't assert this.


> Yeah, well frankly most EDM fucking sucks and that's pretty much why. Your average cool guy "producer" doesn't know the first thing about intervals, counterpoint, or even basics like triads or major/minor scales. But they sure can twist knobs in a way that makes neat sounds!

That's a pretty harsh objective take for a highly subjective topic.

A lot of people like it - myself included. Just because you don't like it doesn't mean it's not valid.


I didn't say I didn't like some EDM. Some is made by very talented people.


At least they have more talent than I've seen in a lot of the newer rap artists, who apparently can't leave home without an auto-tune box handy.

EDM at least has to be considered, and likely, no matter how crap it is, someone put some sort of effort into it. Speaking into a pitch bender takes nothing but breath, and I'm stuck listening to this garbage in traffic every day.


That's not true either. Look at something like Carlie XCX - Pop 2, the autotune is doing interesting things. I'm sure it's not just the technology of it, but also the technique.

Also moving back to one of the most well known autotune users T-Pain is a great singer without the technology. It's obvious there's more to it then just auto-aim function.

Some people use it that way, but it doesn't mean the technology is fundamentally limited to that scope.


To quote from Computer Music by Dodge and Jerse (an excellent reference),

"Helmholtz concluded that the spectral description of a sound had the most straitforward correlation with its timbre.... For instance the qualitative description of 'brilliant' or 'bright' characterizes spectra that have a great deal of energy at high frequencies [e.g. brass instruments] ... A spectrum with little energy in the even-numbered harmonics characterizes the particular timbre that is produced by the clarinet in its low register...." (2nd Ed. Section 2.6)

Understanding that each instrument produces a spectrum of harmonics peculiar to it ... and that those harmonics have relatively different pitches and amplitudes and evolutions ... that vary with how the instrument is played ... is key to quantifying its timbre.

(At first, not an easy thing to grasp let alone quantify. But Partch made a lot of progress!)


I visited conservatoire des arts et metiers museum in Paris a couple years ago. It is full of fascinating scientific devices from the past. Among then was the timbre analyser. I was blown away. A purely analogic spectrum analyser.

https://commons.wikimedia.org/wiki/File:Timbre_analyser-CNAM...


I've gotten a lot of use out of thinking of genre in relation to timbre. Not as the only relevant thing, but part of a constellation.

In most genres you have a palette of timbres to work with for each instrument based on the genre conventions. You can push at the edges or add things in, but have to balance audience expectations carefully as you do it. So for example the way an electric bass sounds in contemporary metal is just not normally going to "fit" into a chicago blues band, even though both are heavily dependent on the sound of electric bass.

Different genres have different relationships to this constraint, for example western classical has a huge palette available in an orchestra but is relatively averse to using instrumentation outside of that collection. Except in percussion, where there is a lot more flexibility! Jazz has a fairly small & rigid set of acceptable timbres for its core instruments, but is fairly tolerant of experimenting with new instrumentation.

So then the two genre-timbre relationships I find most interesting are electronic and pop. Pop is, more than any other genre I think, curious about how timbre effects emotional response in music. An album by the same artist could have a huge range of different sounds for recognizable instruments, using the tension between them for different effects.

And then large swaths electronic music are built largely around active change of timbre over time through a piece. Something you see used conservatively and carefully in most genres becomes almost the central practice.

IDK it's hard to articulate and I'm not trying to set a reductionist framework about how music sounds or anything. Just a line of musical thought I've been exposed to and found valuable.


Well there is Dimmu Borgir, with an orchestra and a choir: https://www.youtube.com/watch?v=El4zsUZjsDc - and it is just nuts. Yes that's 90 minutes, but watch 5 and then decide.

And that's why I love folk-ish metal so much.

Why shouldn't we have a lead/harmony violin on top of a metal rhythm section? https://www.youtube.com/watch?v=F17sxgZdXZw

Or a hurdy gurdy? https://www.youtube.com/watch?v=LT7HqIP55tI

Or a pipa, which was btw very hilarious in poland: https://www.youtube.com/watch?v=MwpKZtPBIEQ

I'm honestly thinking of dropping some money to get myself an A-tuned drone flute[0]. Drone flutes sound really cool, and the drone would be tuned to just about the highest two notes on the guitar... and if you work with it a bit, drums/bass/rhythm guitars helping with rhythms, chords and harmony could really push that somewhere really interesting.

0: https://www.youtube.com/watch?v=j3OlNNEuu88


I went to a interesting hackathon the other day focused on building tools for exploring timbre in sound https://comma.eecs.qmul.ac.uk/timbre-tools-hackathon/.

Was brilliant, a lot of groups focusing on the use of ML to characterise the "unexplainable" in sound synthesis.

We ended up submitting a tool for interacting directly with Ableton using LLM agents after becoming disenchanted with text2audio models, wrote about it here - https://montyanderson.net/writing/synthesis


Very cool. I will have to join "mod wiggler" lol.

I have been out of it so long muff wiggler is now mod wiggler. Come on, that is absurd. Muff Wiggler was the best name ever.


Yeah.... no. Why audio-nerd forums were ever so infantile as to brand themselves with "muff" and "slutz" isn't much of a big mystery, but we haven't lost anything of value by seeing those 'cute' names off. I'd like to chat about synthesizers with fellow nerds without feeling rightfully embarrassed about the name in the header.


Those cutesy names were also a barrier to getting girls and young women involved in this stuff too. Changing it to modwiggler was absolutely the right thing to do.


The lead developer of Melodyne, Peter Neubäcker, has one of the most precise and concrete understandings of timbre. He used that knowledge to create software which changes pitch without changing speed or timbre.

There's a neat documentary floating around which includes some scenes where he explains how he increased his understanding of waveforms in the frequency domain by physcially modeling them in the time domain.


The way a sound evolves in time contains a lot of timbrical information.

Different harmonics have different ADSR curves (Attack, Decay, Sustain, Release).

Above all, one cannot overstress the importance of the attack transient. There are famous experiments in psychoacoustics that show that, when deprived of their attack transient, the sounds of two different instruments may become hard to tell apart.

Personal anecdote: as a classic guitarist, it took me three years of experimentation to find the right way to cut my fingernails in order to have a better sound. The Electronic Engineer in me says that those were three years spent to look for how to improve 0.1 seconds of noise at the start of each of my notes.


>Above all, one cannot overstress the importance of the attack transient.

Roland used this fact to produce surprisingly realistic sounds at low cost. When memory was too expensive to sample the whole note, you could get excellent results from sampling just the transient and using traditional subtractive synthesis for the rest:

https://en.wikipedia.org/wiki/Linear_arithmetic_synthesis


> when deprived of their attack transient, the sounds of two different instruments may become hard to tell apart

A neat trick is playing guitar through a volume pedal. You mute the sound, pluck a string and then swell in the volume pedal. The rest of the ASDR envelope can be created with other effects like delays.


The song Cathedral by Van Halen is a particularly brilliant example of this. Except even more remarkable he used the knob on the guitar to do it rather than a pedal:

https://youtu.be/3OqrWUooOJc?si=NJDoNGkSfJu0gM6X


And reportedly killed the volume pot on the second and final take:

"Eddie did two takes, and the volume knob froze completely at the end of the second take due to the heat generated from rolling it on and off."

https://www.vhnd.com/diver-down/cathedral/

His signature Peavy Wolfgang guitars later featured low-friction pots that are still available for sale as individual components

https://www.vanhalenstore.com/page/VH/GG/EVHPOT


Timbre is what makes the violin sound different from the bassoon, what makes the strings in general gives them their sound, and brass theirs. Timbre is what makes a singers voice unique.

Classical composers had a set of fixed timbres to play with, each instrument having their own. With synthesizers everything is possible.


There are two levels to "everything is possible." First, recorded music is a string of numbers, and any way of generating those numbers will sound like it's intended to. But synthesizing a desired sound is still an evolving process of developing useful interfaces and waveform generators.

Generating that sound in a live performance that is partially or fully improvised is a harder problem.

The evolution of solutions to either problem depends on which instrument you're interested in, and is based partially on Means, Motive, and Opportunity. There's a reason to synthesize a piano or Hammond organ, which is that they're hard to transport and maintain. Less motive for synthesizing a violin.


"everything is possible" except for a synthesized instrument that actually has the expressive range of a real violin or saxophone. Even a synthesized piano is noticeably lifeless compared to the real thing (synthesized piano bass notes are especially thin).


My trombone is lousy at reproducing the sound of a cello, but I don’t consider it inferior.

More to the point, the synthesizer has more options for expressiveness through change in timbre, glissando, microtones, etc.


But nobody made the claim that everything is possible with a trombone.


Is there a synthesizer that claims anything is possible?


For those with the patience, I'd guess that https://en.wikipedia.org/wiki/Wavetable_synthesis would come closest to that ... with the 'right' controllers... and maybe more than one performer!


It depends on what synth you're using, and how you're controlling it. Most synthesizers and controllers are profoundly lacking in expressive control, and of course a synthesizer that just plays back a recorded sample of a real piano isn't going to sound as good as being in a room with the real thing. Just because the mainstream synthesizer world has been sort of stuck in a rut for 40 years doesn't mean this is the best we can do.

There is, however, some real progress being made. For controllers, check out the Linnstrument, the Osmose, the Continuum. MPE was added to the MIDI spec a few years back, which makes it a lot more feasible to sell expressive instruments and have them "just work" with existing synths. As for synthesizers, just about anything is possible. Physical modelling is becoming more popular.


A real piano is also profoundly lacking in expressive control over timbre (excluding extended techniques like half-pedaling or directly touching the strings that hardly ever show up in the repertoire). The only control the piano gives you over note timbre is how hard you press the key. Once you've done that, the hammer disconnects from the key and flies through the air out of your control.

A keyboard with polyphonic aftertouch, such as the Yamaha CS-80 (famously played by Vangelis), is objectively more expressive than a piano.


> The only control the piano gives you over note timbre is how hard you press the key.

This is just not true at all. The action of a piano is a very sophisticated mechanism that conveys both acceleration and velocity, along with the initial position of the key (e.g. fully released or half-pressed). This allows for a huge amount of tonal variation, though it's more subtle than a ribbon controller, and obviously a synth can do many things than a piano cannot.


The piano action physically disconnects the hammer from the key after you press it. Here's a diagram:

https://en.wikipedia.org/wiki/Action_(piano)#/media/File:For...

All the complexity you mention only affects how the repetition lever (part 9) briefly bumps against the hammer shank (part 8). The only information this imparts to the hammer is velocity, and once it happens there's no more contact until the hammer falls back down. The hammer can't somehow remember what caused it to travel at a specific velocity and strike the string differently. There's only a single degree of freedom in its movement. What you perceive as "a huge amount of tonal variation" is mostly note timing.


>This is just not true at all.

It's entirely true for the attack, because of the escapement mechanism. At the point at which the hammer makes contact with the string, the hammer is mechanically disconnected from the key and moving purely through inertia. Once you've tripped the escapement, the hammer is beyond your control and pure physics takes over; the only variable is the velocity of the hammer when the escapement is tripped.

The release phase is controlled by a single parameter, albeit one that can vary over time. The key is holding up the damper, which returns to the string under gravity. A continuous value representing the range of motion of the key would be sufficient to model this with absolute accuracy. Digital pianos invariably sense release velocity rather than key position, because it's much cheaper to implement and is perfectly sufficient in virtually all cases; the range of motion between partial and full damping is so limited as to provide only a very limited degree of expressive control.

https://www.youtube.com/watch?v=XthnCDTnAGw


I don't have a good analytical response to this. Your comment doesn't pass the smell test for the following reason: I am a fairly serious jazz guitarist, and it seems to me like most great guitarists can be quickly identified by ear after playing three notes: Charlie Christian, Joe Pass, Blind Willie Johnson, Mary Halvorson, Django Reinhardt, etc. In some cases just one note is enough: B.B. King, Wes Montgomery, T-Bone Walker, Jimi Hendrix. (60 years later and still nobody plays octaves as well as Wes did).

Obviously some of their sound is tied to specific guitar models, but their tone is still easily identified on other guitars. And although these players were creative improvisers, that's not what I mean. Simply put: if it's a great guitarist, you can hear their hands like you can hear a singer's mouth. (Though hands are clumsy compared to mouths, and guitars are clumsy compared to horns.)

I am not anti-synth or anti-electronic music. But I do not believe any existing synth controller provides the expressive possibilities of an electric guitar, even Allan Holdsworth's goofy breath controller. Maybe I'm wrong and there's a great example out there on YouTube. But it seems to me that for an individual to express their individuality with a synth, they need to play a longer musical fragment and color their expression with timing, volume, vibrato, etc. Three notes "with pizazz" will not really be enough.


There are a lot of famous guitar players, but there really aren't a lot of expressive controller players. I don't know of anyone who is famous for playing the Linnstrument, or the Osmose, or Continuum.

Vangelis and the CS-80 is maybe an exception, though even if his songs are very recognizable, probably a lot of people could convincingly imitate him.

Maybe Wendy Carlos could count too.


Carlos was great at deliberate performances, but if there are any 'Carlos Live' recordings, I haven't heard one! But then LOTS of people grew up learning to play instruments from great teachers for centuries ... not so many with synths.


Only the the best grand pianos have any bass. Your cheap 43" tall console piano for apartment use will not have much bass to speak off. The strings are too short, the box is too small. We have to restrict ourselves to True Scotsman's pianos.

Bass is easy to obtain out of a synthesizer: just make sure you have a subwoofer in the speaker system. It doesn't have to be a True Scotsman's synth or subwoofer.


I am both a classical guitar player and long time synth head.

"Everything is possible" is a bug and not a feature in many ways.

It certainly is possible to synthesize an expressive instrument to the level you describe but what usually happens is you move on to something else.

"What does this knob do??" Wahahahaha

That kind of thing is just so much fun compared the guitar that you have to actually learn to play.



I've always thought of Jimi Hendrix as an example of timbre used for revolutionary purpose. A lot of what he acheived he acheived with timbre.


He definitely opened the door for timbral exploration on the guitar using electronics.


> A vocal timbre can also reveal economic and cultural background. Some singers...were encouraged to either accentuate their timbre...or to abandon what occurred naturally for them

Vocal timbre is a big identifier of different rap subgenres and artists, seems like. Certainly seems like much of the vocal delivery on a track explores fluctuation of the timbre, even when the pitch is constant or has little dynamics.

It's interesting to think how the timbral effects of a regional accent would interact with the beat construction/development. Is the trap hi-hat a result of the Southern drawl?


The reason I hate heavy autotune in recent hip-hop is because it destroys all of the interesting timbre that could be there, and reduces it all the same annoying voice


Depends on the usage for me; I think the point of the article was that timbre isn't something that's orderable, there's lots of different timbres and processed vocals are their own thing. Sometimes their texture fits really well into the palette of the song, IM personal O. Thinking of like "Lollipop"


As a synth enthusiast I generally think of timbre in relation to "real" instruments, brass, strings, piano, etc. In the article they mention there are as many as 10 dimensions but don't actually define what those are. Timbre is something I've thought about a lot but nailing down a more general classification seems pretty challenging.


That’s funny, I’m more likely to describe “real” instrument timbres in synth terms, many of which I bet relate closely to the 10 dimensions. Oscillator waveform, envelopes, filter shape/cutoff, resonance…


Synth players have studied and emulated timbres (and created new ones) since the pipe organ (an additive synthesizer). Many books have been written about timbre in electronic terms -- a combination of frequencies and amplitude envelopes that either sound to our ears like classical instruments, or like something else entirely.

I would go so far as to say understanding and selection of timbre is 50% of the work of a synth player.


Tangentially, Sennheiser HD600 headphones are known as "timbre kings" for their ability to reproduce timbre accurately.


Pitch is the frequency of the oscillation, timbre is the internal structure of what's oscillating. Is that right at all?


"Everything is overtones", or so I've heard somewhere. The pitch, or note you hear from an instrument is the fundamental frequency, and the combination of overtones determines the timbre. These overtones are other sine waves at different amplitudes and frequencies, and the combination of these are instrument dependent. Additive synthesizers do this to approximate "real" instruments.


The combination of overtones, but also the change in volume of those overtones over time.


A thousand times this.


It's also the shape of the wave. Sawtooth waves are buzzier or "brighter" than sine waves, for example.


And the shape of the wave is just a composition of more waves :->


This is a great visualization of fourier series.

https://www.youtube.com/watch?v=r18Gi8lSkfM


Man I wish my full time job was studying math.


For periodic signals the waveform and the overtone series precisely determine one another.


And the human ear is mostly insensitive to phase of those overtones, so you can have waves with visually very different waveform that sound identical.


Timbre is the quality that's left after you factor out the pitch and intensity. How it "sounds".


Yes, that sounds mostly right to my understanding. Exploring the space of that internal structure is the point of the article. Both pitch and timbre get remarkably complicated the more they're looked in to.


Timbre can vary a lot day by day on the same instrument.

Violin and the bow are both made of wood and a rainy damp day can easily sound different than dry summer's day.

You can get good carbon fiber bows and they are much less weather sensitive.

You can also get carbon fiber violins but I haven't heard one that sounds as good as one made from wood.


We can get closer to the mathematical definition of timbre by talking about the strength of the harmonics https://www.youtube.com/watch?v=VRAXK4QKJ1Q


A fine article on the role of timbre which is indeed discussed little compared to the focus on harmony and form for classical music. Charles Rosen gave one possible reason for this neglect in his book "Piano Notes", where he traces it to a philosophical prejudice of composers from Hadyn onwards against variety in tone colour and in favour of more abstract qualities.

> The utility of the piano for composing was its neutral and uniform tone color: in theory (although not in reality) the tone quality of the bass is the same as the treble. In any case, the change in tone color over the whole range of the piano is, or should be, gradual and continuous (there are breaks, of course, when the notes go from one string in the bass to two and then to three in the treble). The monochrome piano might be used therefore just for its arrangements of pitches, and the quality of the sound could-absurdly in many cases-be considered secondary.

> What made it possible for composers to refuse to acknowledge the difference on the piano between treble and bass and leave whatever problems arose to be solved by the performer was the fact that the change in tone color over the span of the keyboard is not like the leap from a bassoon to a flute but continuous and very gradual when the instrument is properly voiced. These imperceptible gradations are the result of a deliberate policy of a unified sonority on the part of musicians and instrument makers. All attempts over the history of piano construction to incorporate anything analogous to the picturesque changes of registration in the organ and the harpsichord had little success, were not exploited by composers, and were finally abandoned. Radical contrasts of tone color were traded for the possibility of making a gradual crescendo or diminuendo. This was a decision that took place at the same time as the preeminence accorded to the string quartet over all other forms of chamber music; that, too, emphasized the importance of a unified tone color. Chamber music with wind instruments, while the occasion for several masterpieces, became the exception, an exotic form. That is why the use of colorful sonorities in the orchestra has so often been considered somewhat vulgar, as if calling attention to the sound were paradoxically to detract from the music. The prestige given to pure string sonority is part of the asceticism of nineteenth-century high culture. Contrasts of tone color were given a significantly lower place in the hierarchy of musical elements. This is one reason that only the piano repertory rivals the string quartet as the most respectable medium for private and semiprivate music-making from Haydn to Brahms.


Highly recommend the book “Tuning Timbre Spectrum Scale” for engineers interested in music theory

https://sethares.engr.wisc.edu/ttss.html


Arnold Schoenberg observed the deep connection between music, perception, and cognition; anticipated composers applying timbral control as is now commonly done with synthesizers; and apparently did not anticipate non-subtlety of the effect in EDM:

"I think the tone becomes perceptible by virtue of tone color [timbre], of which one dimension is pitch. ... Pitch is nothing else but tone color measured in one direction. Now, if it is possible to create patterns out of tone colors that are differentiated according to pitch, patterns we call 'melodies', progressions, whose coherence evokes an effect analogous to thought processes, then it must also be possible to make such progressions out of the tone colors of the other dimension, out of that which we call simply 'tone color', progressions whose relations with one another work with a kind of logic entirely equivalent to that logic which satisfies us in the melody of pitches. That has the appearance of a futuristic fantasy and is probably just that. But it is one which, I firmly believe, will be realized.

...

Tone-color melodies! How acute the senses that would be able to perceive them! How high the development of spirit that could find pleasure in such subtle things!"

-- Arnold Schoenberg, Theory of Harmony, 1911, toward the end of chapter 22.

Personally, I anticipate that if humanlike AGI is achieved, then it will understand the illusion of music. Not merely because music understanding is an attribute of humanlike, but because there must be something intrinsic in the process of human cognition that causes the illusion of music occur. Music understanding could be viewed as an essential waypoint on the path to AGI.


I had a simple thought a couple years ago that still fascinates me. If you do something as simple as "clap clap clap" (with equal measures of time), you can naturally relate it to any other set of sounds with that simple pattern (ABABA). moreover, there is a 1 to 1 relation of this abstraction in space, where one can represent the sounds as marks on a screen (as I have done here for example, as "clap clap clap, X X X, - - -, or whatever).

Despite the different sensory systems (sound, or visual) and the difference in actual data (I don't know how many photons or audio waves entering your brain), we can abstract them down and understand/relate them with a simple pattern.


I think you would enjoy reading James Gleick's The Information!

Edit0: https://en.m.wikipedia.org/wiki/Talking_drum

It begins by talking about this, maybe that's a good start! It documents that they are able to transmit complex messages like jokes and proverbs entirely through rhythm!


Thanks! checking it out now


What you illustrate is deeply fascinating. In music, there is a grammatical perception that works both cognitively and as an embodied experience.


Yes, but its not just music. Our senses are very much linked in that we can build a one to one relation of ideas in visual space to auditory space and vice versa. There is a common abstract model we understand so to speak.

Our model of perception is very finely tuned with that of space-time.


Pop music really does do this as he described. There are lots of songs that play with expectations of timbre in the same way that normal music plays with expectations of pitch.

When I get mentally crystallized, I hear music as proof-of-work in the vein of Bitcoin hashes.


I very much appreciate your comment, and at the same time am interested to challenge this idea of "the illusion of music". What about music is an illusion? Maybe this was just an unfortunate phrasing, and yet it might touch on the heart of the matter. Music is a language of relationships, indeed it is nothing but relationships, in time, pitch, timbre, volume, voice, number and so on, non-symbolic and changing relationships of sound carrying aesthetic and emotional feeling and meaning. It is both objective and subjective -- objective in the sense that "4 + 4 = 8" is an undeniably more beautiful statement than "4 + 4 = 6", and subjective in the sense that you and I may have different favorite numbers and also we may react differently to various medications or prefer different foods. Music is external and internal, social and individual. What it is not ever is an illusion as then it would not exist. (It may express illusion, or become more like a creator of illusions, for example when the orchestra tries to conjure up the illusion of a thunder storm.) Now, one might be tempted to counter that the kind of "meaning" music carries is itself an illusion, an epiphenomenon maybe, but then why should one care about anything at all, including AGI, since the kind of meaning that music carries is the same as the meaning of love, of hate, of good and bad, pains and pleasures, and hopes and regrets, tragedies and successes?

Maybe a reason for calling music an illusion is to try to point out that it's all in our heads. Sure, the physical instruments are over there, but they are just some atoms and molecules rattling about creating vibrations that, like everything ultimately, share the meaninglessness of the physical universe of just more material stuff (ignoring the quantum and the unknown). In which case though life seems also to look more like an illusion. Ah but the meaningless vibrations enter our ears and are processed and become music, so like color there is none but in our heads, thus illusion? But these notions also fall apart. The redness of apples isn't all in our heads; it requires a certain relationship to light frequencies that we see as redness. It's not an illusion it's just what red looks like through a living human body. If not through our senses and brain, and then mind, where should we experience it?

Music is only ever meaningless if it is unheard, as apples can only not be red when they are unseen (or are Granny Smiths).

So music is only meaningless if there is nobody -- no body -- to listen. And maybe that's where all of this leads. Your AGI needs to be embodied and alive.


My use of illusion was not intended to diminish the significance of music. To the contrary, I meant to distinguish it as an exceptional perceptual phenomenon which reveals important aspects of the inner workings of human experience.

Somehow, the physical situations of our bodies become perceptual experiences. Usually, those perceptions are mere descriptions of the physical world, for physically navigating it. Some exceptions exist, such as music and optical illusions. The translation from air pressure fluctuations to a tree branch falling is one type of experience, the link to love or hope is another. Rather than being insignificant, my sense is that music understanding will prove to be an emergent capability available in whatever architecture might produce humanlike AGI, because it connects human perception and cognition too thoroughly to not be baked-in. Going further, building models to understand music would be an excellent step toward embodied AI with humanlike cognition.


This is something I've been thinking about the last 3-4 years and desperately what I am trying to do.

I have what I think some interesting ideas and implementations, but no one quite grasps my thought process and I am stuck between believing Im onto something and that I am furiously chasing nothing.


My experience in pursuing this is that I applied to about a dozen on-campus CS grad programs, over the course of two years, noting this research direction, but those applications were all rejected. You would think with the emergence of generative AI that there could be some interest. I suspect the application was of interest, but that universities have a shortage of experts in generative AI, and that a multidisciplinary project like this might be misunderstood within a CS department or would have to be accepted by a very specific advisor who may not exist. The best chance might be to convince investors to help assemble the right team and to provide a viable R&D budget.


Oh wow thats nuts. Did you end up attending grad school for something else? Ive been thinking about applying but I haven't even able to find any departments that do research in this area. There is a huge craze around ML as its producing tangible results and it seems the school of thought is that anything else is wasted time.

The closest thing Ive found to my own is Wolframs writings.

Seriously I know I seem like a delusional nut but this train of thought has really grabbed me. Ive written hundreds of pages of (mostly intelligible) notes/figures and spend countless nights trying to implement what I can so clearly see in my head.

Unfortunately I dont have anything concrete yet that would warrant building a team, and its hard to even find people I can discuss my ideas with.

I think there ought to be interest in this area in mathematics, philosophy or even linguists though.


Would you mind reaching out through a contact method in my profile? The discussion is getting less general so that might be a better avenue.


This is tangential to much of what this article discusses, but I've had this idea kicking around in my head for a while, that, to some degree, the artificial imitation of acoustic instruments using synthesizers is to music what skeuomorphism is to design, and I (generally) dislike it for all the same reasons. This isn't to diss synthesizers on the whole - I unequivocally love them, and they really opened up my musical world by a wide margin. Nor am I saying that there is no place for synth patches that attempt to "sound like" "real" instruments (for lack of a better word; I consider every "model" of synthesizer to be its own very real and distinct instrument).

Skeuomorphic design, to me, largely feels lazy and unattractive; rather than designing for the new, maybe unfamiliar, medium you're working in, you attempt some facsimile which inevitably cannot live up to the original, either out of lack of ambition, or lack of faith in your audience to understand or appreciate without the anchor of a common metaphor. It lacks idiomaticity - there are particular details, quirks, associated with different mediums that lend themselves to different sorts of designs, and a skeuomorphic design language ignores these. (The same is true of musical instruments. A piece written for the lute, say, played on piano, would feel very different than a piece written for piano, played on piano; what is easy, or possible, on one instrument, is not necessarily as sonorous on another.)

To me, the use of synthesizers to emulate acoustic instruments, where the express intent of the composer/producer/whatever is to evoke the sound of that acoustic original (read: where the creator would prefer the sound of the real instrument), is, at its most generous, telling of a lack of creativity and at worst laziness. (If what you want is expressly the sound of a synthetic imitation for the purposes of your art, none of this applies.) It is a lack of creativity in the case where the creator simply does not have access to "real" instruments they'd rather use, and it's laziness where you have access and ability but opt for something you yourself deem inferior.

Of course, this all presupposes that there is something lost when say, a synthesizer plays a violin patch in an earnest effort by the creator to emulate a violin - where a real violin is actually what's desired. I think this isn't controversial to say, but of course it's a spectrum; I have much less of a problem in cases where the delta is smaller; e.g., simple legato harmonies from a string section can be emulated more convincingly (by orders of magnitude) than some virtuosic cadenza by a solo instrumentalist.


At least this article will reinforce the concept, and the correct spelling, if not any appreciation for the quality!


Everything we can't describe in music" is a thought-provoking statement that highlights the limitations and strengths of music as a medium.


"Terroir" is a terrible way to metaphorize timbre. The only similarity is that they're both hard to describe. Terroir is there before the wine is made and can't really be controlled by the winemaker (except insofar as they choose where to plant or what grapes to blend). Timbre is absolutely under the control of the musician.


> Terroir is there before the wine is made and can't really be controlled by the winemaker (except insofar as they choose where to plant or what grapes to blend).

I've met winemakers who think of planting in terms of multigenerational experiments with highly intentional controls for different aspects of soil and micro-climate that may vary on a scale of a single meter, who work to find and train their successors because they will not see the results of some of these experiments in their lifetimes; winemakers who absolutely understand that blending is not just a linear combination of the characteristics of blended wines, but a truly complicated system where certain flavors may heighten others; winemakers who rely on centuries of experience in how funguses and other microorganisms can be used in safe and flavor-enhancing ways.

Winemaking is engineering, just as musical instrument design is (both physical and digital), and just as building a startup is. The presence of human subjectivity as an evaluating oracle in a complex system does not imply that its creators work without intention or without rigorous systems-level thinking.


As a string player, terroir strikes me as a beautifully appropriate way to describe timbre. My instrument was constructed in the 19th century, probably in Germany. The luthier, their choice of wood, the shape, the workmanship, the strings, the bow, the weather, the player, the room, the mood, the other players all affect the sound. A large portion of the job of an instrumentalist is choosing which tools to use. There are always compromises, and many variables that effect the timbre simply are before a note is played, and have to be worked with.


I think the sense is more they're both about hundreds of unique physical characteristics, some of which are perhaps unquantifiable.

The shape of an instrument, the material used, the type of strings, their tension, the temperature and humidity in the room. The way each player holds it, the size of their hands, the specific strength of their fingers. Some of these factors stay constant, others change from day to day or player to player. The sum of all of these together creates a specific timbre.


Thanks, that analogy bugged me a lot and I couldn't get why a farmable region was similar to a variable sound envelope.

A terroir is totally describable, as by definition you have uniform geological and climatic properties (on the other hand if it can't be delimited it's not a terroir)


I think this is just another point of frustration with trying to define timbre. There's definitely some points directly under control: attack and brightness come to mind. Then there's some that are less directly under control: the room and the instrument itself.

It's no accident that musicians spend a lot of money on instruments -- even electric guitar players -- an instrument that is electronically amplified and often distorted and otherwise augmented -- will have strong feelings about the particular type of guitar they use. Leaving shape-based decisions out of it (and pickups), hollow bodies are going to produce a different sound than solid. Different wood types will produce different tonal characteristics.

So the things that are directly controlled are less like terroir, more like choice of grape used maybe? While the make of the instrument, the shape of the room a recording was made in, these are much more like terroir I think.

But it's probably just not a great metaphor.


>Different wood types will produce different tonal characteristics.

True for acoustic guitars, but for electric guitars this is thoroughly debunked:

https://www.youtube.com/watch?v=n02tImce3AE

Previous HN discussion:

https://news.ycombinator.com/item?id=36106674


TIL! I grew up hearing about it so much from the broader guitar community that I just took it as truth.

The idea holds across many acoustic instruments though.


> Another way to think about timbre is by comparing instruments. In an old stand-up bit, Steve Martin strummed the banjo and mused, “You just can’t sing a depressing song when you’re playing the banjo . . . You can’t just go, ‘Oh, death, and grief, and sorrow, and murder.’

Anyone who's played Outer Wilds should disagree with this.[0]

To say we can't describe timbre is a bit misleading, because there are concrete mathematical ways to analyze sounds; they exist in a three-dimensional space of frequency, amplitude and time. But that's helpful in the same way that describing programming languages as collections of 1s and 0s is.

What's lacking for describing timbre, I suppose, are the steps between "this sound is a sum of a particular arrangement of sine waves" and "this sound is a piano". There are common terms such as ADSR or "brightness" and "warmth" but those don't tell the full story.

The question is, how valuable is that intermediate step when you could just say "this is a piano, this is a banjo"?

[0] https://www.youtube.com/watch?v=YR_wIb_n4ZU


A banjo is plinky-plonky: it has a sharp attack, but not a lot of sustain. It's not great for playing a melody with long, slow notes. It's better suited for rhythmical patterns, and faster tempos. It's possible that a banjo could have a part in an ensemble piece about death, grief, sorrow and murder, but as a solo work, the idea does seem dodgy.

It's not so much the timbre as the envelope. We have ways to talk about the not envelope. That is complex though, because different components of the sound can have different envelopes: e.g. high harmonics dying off before fundamentals. Envelope is not separable from timbre.


I'd also recommend Timber Zeal - https://www.youtube.com/@SpaceBanjoMusic https://timbrzeal.bandcamp.com/music ... and the Hardspace Shipbreaker opening https://youtu.be/b4LbAr4uz9A

You certainly can do minor with a banjo.


Notably easy when you tune the B string to a B flat. Instant minor!


Thank you VERY much!


You might also find Hardspace: Shipbreaker - Music of Space https://youtu.be/RX_56MiOnAk interesting.

> Watch our audio director Ben McCullough describes how he created a unique audio environment and soundtrack for a Hardspace: Shipbreaker.


I always thought the Outer Wilds OST was rather hopeful, in a rustic kind of way, not depressing.


I don't know if I necessarily agree with that quote but your counter example does not give me emotions of death, grief, sorrow, or murder. It sounds very hopeful to me.


And that just goes to show you how context-sensitive and subjective the emotional quality of sounds and music is. Half the people watching that video get teary-eyed when hearing it :)


> Anyone who's played Outer Wilds should disagree with this.

Not exactly about death, but certainly grieving things like the emotional blunting of psychiatric medications, inability to accept love or forgiveness, and lines like:

> I'll pretend bein' with you doesn't feel like drowning

Sounds decidedly less hopeful to me than that track from Outer Wilds

live: https://www.youtube.com/watch?v=psIRw0d509w

studio version: https://www.youtube.com/watch?v=G-XICfi4j3Q


>You can’t just go, ‘Oh, death, and grief, and sorrow, and murder.’

There's a whole sub-genre about this that's commonly played on banjo:

https://en.wikipedia.org/wiki/Murder_ballad


Is that the space fallout game? Can't say I remember the music.

Edit: I was thinking of The Outer Worlds, not Outer Wilds btw


Tone is in the fingers, as they say for guitarists. Different woods, magnets, strings all impact tone for sure, but at the end of the day Jimmy Page sounds like Jimmy Page and Hendrix sounds like Hendrix cause tone is in the fingers.

That’s what stops music from being math.


This story was not submitted by tintinnabula.

I'm scandalized. /s


tldr: Timbre, the unique quality of a sound that distinguishes different voices or instruments from one another, is complex to define and measure and is often overlooked in music theory, which tends to focus more on pitch, rhythm, and harmony.




Consider applying for YC's W25 batch! Applications are open till Nov 12.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: