Fall Sale: 15% off at checkout!

On Audio Fidelity: The Concept of Fidelity

Part 1 of 3

“My God,” I cried in horror, “what are you doing, Mozart? Are you serious about inflicting this monstrosity upon yourself, and upon me? About subjecting us to this hideous contraption, this triumph of our time, the final, victorious weapon that our age deploys in its war of annihilation against art?”…

Thomas Edison with an early twentieth century phonograph

He regarded my agonies with deep pleasure, he turned the cursed screws, adjusted the metal cone. Laughing, he continued to let the distorted, disfigured, and poisoned music seep into the room; laughing he gave me my answer.

“No pathos, please, dear neighbor! By the way, did you notice the ritardando there? A stroke of genius, no? Ah yes, and now, you impatient man, let the spirit of that ritardando sink in—do you hear the basses? They stride along like gods—and let old Handel’s stroke of genius penetrate and calm your restless heart! Listen without pathos and without mockery, my little man, and behind the veil of this ridiculous contraption—which really is hopelessly idiotic—you will hear the distant figure of that godly music passing by!…Observe this absurd metal cone, it seems to be doing the stupidest, most useless, most forbidden thing in the world, just taking music that’s being played in one place and indiscriminately blasting it into another place where it doesn’t belong—stupidly, crudely, and miserably distorted, too—and nevertheless it can’t destroy the original spirit of that music, all it can do is use that music to demonstrate the helplessness of its own technology, the mindlessness of its mechanical operations!”

—Herman Hesse, Steppenwolf, 1929.

Audio fidelity is a fantasy, not a reality. But it is a fantasy that can show us true things about the world—and like all fantasies, it contains a utopian impulse, a dream about hearing that manifests in an unfolding future.

The idea of “fidelity” divides audio enthusiasts into roughly two camps, both of which blend together large portions of unfounded ideology, subjective impressions, and pieces of our best current understandings of sound, electronics, mathematics, psychology, and sound reproduction. Efforts by both camps have at certain times improved our understanding of sound, hearing, and electronics, and at other times have impeded progress through attachment to ideas which later were found to be suspect. On the one side, there are the scientific skeptics, who generally work with measurements taken by nonhuman instruments and derived from basic engineering principles. Although the analogy is far from perfect, let’s call this Camp Skully, in reference to the character from The X-Files. On the other side, there is Camp Mulder, which works empirically, trusting subjective impressions and distrusting measurements. While this description of the two camps signals real differences in approach, like any opposed ideologies each camp is often defined more by its opinion of the other camp than by its own beliefs. To Camp Skully, Camp Mulder is at best based on demonstrably unreliable subjective impressions, and at worst full of con men and rubes. To Camp Mulder, on the one hand Camp Skully contradicts the very idea that we can experience things and that our experiences are real, and on the other hand views science as a fixed body of knowledge instead of an unfolding process of knowledge production. Of course, neither side is entirely what their opponents believe they are, and neither side is entirely what they believe themselves to be.

Mulder and Skully

As electronic musicians, these debates are not addressed to what we do, and yet they nevertheless surround our musical practice. For example, while digital recording is not so different from analog recording, digital synthesis often requires entirely different mathematics that will react very differently in a complex patch. And yet, searching for information on digital audio, we will doubtless encounter mostly debates about the merits of digital and analog in the faithful reproduction of sound, with perhaps a few minor debates about analog and digital mixers that are sort of relevant to us. In many ways, our problem as musicians is that the fidelity controversies produce enough text to overwhelm and heavily color any practical information we might need. But while these controversies are a problem for us, they nevertheless also have things to teach us.

This exposition is divided into three parts. This part will cover the big ideas of fidelity, detailing the usefulness of the idea, but largely focusing on its problems and limits. The next part will deal with how fidelity is measured, and what these measurements mean. The last will focus on the characteristics of digital and analog media.

The Concept of Fidelity

Part of the reason for such wild controversy around fidelity is the inherent difficulty of defining the concept. From its earliest articulations, philosophy has been troubled by the way that two different objects can seem to be copies of the same sort of thing. We are inundated by examples of striking similarity: two ingots of copper, two sparrows, two pronunciations of the same word. If these were identical in every respect, they could not actually be two different objects in the first place. At the very least, they differ in location, time, matter, and relation, but practically there are always many more differences. Platonic and Aristotelian philosophy thinks about this by imagining some perfect, single, conceptual form for a type of thing, shaping many objects like a single brass stamp can shape many wax seals. Today, nobody much likes this weird concept of perfect forms, but its not as though we have solutions that are so much better. Physics has very little to say about similarity or sameness, focusing instead on mathematical relationships between measurements. Aristotle made use of the concept of a form in defining and categorizing animal species—“species” is in fact the Latin translation of the Greek word for form in Aristotle’s works. But modern Biology prefers to organize the tree of life by genetic relationships, rather than similarity.

Practically, there are countless easily measurable differences between any sound and the best possible reproduction we can make. Some of these differences are subtle and small, and others are big and obvious. Discussions about fidelity deal with these differences through the idea of the difference that doesn’t make a difference, differences that are less or more important, not important at all, not “part of” sound at all, differences that it is silly to care about, etc. Camp Mulder often hangs on to some (nonexistent) absolute fidelity as a limit we must constantly approach. Camp Skully is more likely to define specific measurements and thresholds which together define “good enough” or “indistinguishable” fidelity. But to make any sense at all of the idea of fidelity, both camps must adopt this idea of a difference that doesn’t make a difference, and a difference that does make a difference.

For a musician, the problems with the concept of fidelity start right here at the beginning. First, when we are actively creating things, what is this original from which we are supposed to measure differences? Second, why should any difference not make a difference? Why should we not care, deeply, about things that others find trivial? One of the many functions of art is to foreground and highlight differences that others have passed over and forgotten. Why, for example, shouldn’t it matter whether your instrument is a material object or only virtual software? The production and materiality of the art object has been an explicit, central concern for art and music for at least the last hundred years or so. Hendrix set his guitar on fire, but you cannot burn a software plugin. Maybe it doesn’t make a difference, and then maybe it does. Part of what we must do is choose the differences we care about. And of course, it could just as easily be an artistic goal to create a sense of separation from materiality that is only possible with software.

Part of what’s at stake here—for all sides, Mulders, Skullys, and musicians—is the practice of modeling and its relationship to art and reality. Models are so fundamental to science that one might even say that scientific knowledge is models. Camp Skully thinks so, although they think observations are also important. Camp Mulder might think that our observations are the real knowledge, while models are just things we say. The worst disagreements about fidelity usually come when someone from Camp Mulder explains their observations with bad models in which they are not actually that invested, while someone from Camp Skully dismisses the Mulderite’s entire subjective experience just because their models don’t make any sense. For musicians, the problem is that these models are at once descriptions of how our tools work and assertions about what sound and music are. While the former is something we need in order to understand our instruments, the latter is unnecessarily constraining, particularly as debates about fidelity and studies on fidelity rarely move past a nineteenth century concept of music, ignoring over a hundred years worth of thought about chance, space, synaesthesia, movement, noise, context, events, audience, subjectivity, spirituality, materiality, etc.

Models and understanding are a vital part of our work, but beware when Camp Skully “explains” what sound is, what music is, and how silly we are to care about this or that; beware all the more the bad models of Camp Mulder, or its occasional suggestion to go entirely without models. For musicians, the difficult philosophical issues of the relationship between modeled understanding and subjective experience often evaporate, replaced with a practical requirement to learn how to use a particular device as best as possible, along with an artistic impulse that was never negotiable.

The Model of Fidelity

A man sits in a low
chair in front of speakers that sit on his floor. His hair, necktie, and the
lampshade are being blown backward.
Maxell tape advertisement, 1983. The subjective experience is foregrounded. Also note that, as exciting as this is, the listener remains perfectly still. Guy Debord would have a field day.

For both camps, the model of fidelity begins with the subjectivity of a “normal” human listener. This listener has two ears, and each ear receives a one dimensional stream of sound, which consists of vibrations in the air lying within the 20Hz–20kHz range. The ear converts each one dimensional stream into a spectrum, a perception of the various loudnesses of different frequencies. The mind then collects related frequencies into particular, individual sounds, primarily by matching frequencies to harmonics. Next, it compares the relative loudnesses of these frequencies in our two ears in order to construct an imaginary two dimensional image of sound in space. Lastly, the mind observes how these sounds make reverberations in order to construct a map of the space in which the sounds are happening. This listener is stationary, looking at the sound as if they were looking out the window, separate from this sound, neither actively exploring the space nor participating in the creation of sound.

The sound being heard is abstracted from the medium that actually produces it—the listener’s audio system and the space in which it operates. The overall experience is then judged. Classically, any difference between the listener’s experience and the experience of someone sitting in the imagined original space would be coloration or distortion, negatively impacting the fidelity. Recently some people (largely in Camp Mulder) have argued that the overall quality of the experience and perhaps the “sense” of realism or presence can actually be enhanced by intentionally and subtly altering the original sound. However, even in that minority, virtually nobody can agree on what kinds of changes ought to be made.

For all the talk of “objective” and “subjective,” the central disagreement of these two camps is really just whether we should trust people to know their subjective experience or whether we should trust scientific researchers to know about people’s subjective experiences. Both sides do not consider an objective, anti-humanist approach to sound, an approach that would incorporate existence and possibility, nonhuman, disabled, or alternate spaces of perception, or the capacity of subjects to interact with and change the world, instead of simply observing it.

In Approaches to Sound, I already detailed some of the potential problems and benefits that come with the psychoacoustic approach. Here, I will address only a few common models and practices from the fidelity discussions. Like in most scientific practices, as the findings of psychoacoustics, acoustics, electronics, and mathematics make their way to non-researchers, the nuance of what is well-known and what is only suspected tends to disappear, as do variations and margins of error.

Phase and Space

For the most part, although the fidelity model is limited and inexact (and this limitation is also what makes it useful), it is not exactly wrong. There is one aspect, however, which is problematic enough that some more nuanced models of fidelity are modified to account for it: the complete absence of phase. For over a century, it has been understood that humans make use of phase information to understand the position and spatiality of sounds. We find a sound’s position based not just on relative loudness, but on which ear receives a sound first. However, our sound reproduction systems essentially have no hope of accurately reproducing this effect. While spatialization is a topic unto itself, this section summarizes the central tenets of spatialization in stereo audio, and explains why most people think that amplitude, rather than phase, provides spatial cues.

A sound source will have different distances to each ear, resulting in different phases.

In the early twentieth century, Lord Rayleigh developed a psychoacoustic understanding of the way in which human hearing locates a sound source in space, which was slightly revised and expanded by others’ work on localization in the 1920s and 30s. A point source emitting a sound will present different signals to each ear. With the different distances to each ear, it will take a different amount of time for the sound to travel, and also the sound might have to get around your head to reach one ear, while being able to directly reach the other ear. Roughly, we can divide human spatial perception into three frequency regions. At higher frequencies, with wavelengths nearing the size of one’s head, the head will function effectively as a filter, altering the relative amplitude of the sound in each ear. Conversely, the filtering effect, as well as the small wavelength, means that the phase difference between one ear and the other will be affected and might not give good information as to the location of a sound. At lower frequencies (below around 700Hz), with wavelengths lower than the size of one’s head, not much filtering will occur, and the two ears will hear with equal loudness. However, the phase differences resulting from the different distance between the source and each ear will be preserved, allowing us to infer the direction of the sound. As the frequency decreases, the same interaural delay will result in a smaller phase difference, and while experiments have shown an ability to determine direction down to the bottom of the range of hearing, this becomes more difficult for frequencies below 60Hz or so. Lastly, very short sounds will be located not according to a phase difference per se, but according to the time difference between the arrival of the sound at each ear.

Because the ears hear phase from both speakers, this phase difference more or less cancels out completely.

Early attempts at spatial audio tried to use multiple speakers to reproduce the actual structure of a three-dimensional wall of sound moving toward the listener. This would be a more objective approach to spatial audio. But needless to say, this was both difficult and expensive. The much simpler method was just to reproduce what the listener would have heard, the two signals to each ear. While this was somewhat possible with headphones, with two speakers the situation seemed pretty hopeless. Signals from both speakers would of course reach both ears, and these overlapping signals would add together to eliminate the recorded phase differences. The problem seemed intractable.

In 1931, Alan Blumlein created a system to address these difficiencies, creating the modern concept of stereo. He realized that we can recreate the required phase differences from scratch by combining sounds in the medium of the air itself. Because the distance from each speaker to each ear is different, the sound from each speaker arrives at each ear with a different phase. As we saw in Phase, Delay, and Mixing, two sine waves with different phases and different amplitudes will combine to form a single wave with an intermediate phase that depends on the relative amplitudes coming from each speaker. If the waves coming from the left and right speakers have different amplitudes, they will combine to create a different phase for each ear, as if the sound came from somewhere between the speakers.

Although this illusion is convincing, the exact pair of signals we would have heard from a single sound source is not generally reproduced. Panning with volume creates a phase difference in our ears, but also creates a volume difference that does not necessarily reproduce what happens when a single sound travels through the air. (This is partly why there is such controversy about pan laws, the way panning changes the relative volume of the left and right channels—there is not actually a correct way to do it.) Further, for the most part we locate very short sounds (clicks) based on the time difference in the arrival of the initial wavefront. Relative volume will do nothing to modify the initial arrival of each wave.

Headphones are much better at controlling the phase information received by each ear, but unlike speakers they do not convert relative volume into relative phase. There are algorithms called head related transfer functions (HRTFs) that can be used to spatialize sounds for headphones, but they are only rarely used in music production. Nor do a mixer’s panning controls alter phase. When a recording produced for speakers is played back on headphones, we receive spatial cues in a strange, otherworldly way. Our ears receive differences of phase and volume that would not be produced by sounds travelling through the air. Conversely, a recording produced for headphones will lose essential phase information and feel flat when played back on speakers. Ideally, we would produce entirely different mixes for these two media. Given listening practices, nearly all modern mixes are compromises between these (and other) media, while older mixes often translate poorly to headphones.

Confusion about the relative nature of phase adds to the perception that phase does not matter. Absolute phase does not really exist, only the relative phase between two different waves. Consequently, of course absolute phase is not perceptible, while absolute frequency and amplitude (which do exist) are perceptible. In the stereo model, phase differences between the left and right channel are more or less eliminated, and thus our speakers have already destroyed what we would otherwise perceive. The only other way phase exists in these recordings is in phase coherence, the relative phase of different frequencies in a harmonic sound. Bleeding edge research indicates that phase coherence might be used to detect the closeness of a sound, or whether a sound is a reflection or a direct source.

For all these reasons, the stereo model leads to the misconception that phase is not perceptible, or if phase is perceptible, this is subtle and only the most capable listeners will notice. In fact, so far as we can tell, all human binaural listeners generally make use of relative phase information in discerning the positions of sound in space. It is just that stereo speakers are very bad at properly reproducing recorded phase. Fortunately, while most discussions of fidelity don’t fully understand these issues, some do, and others are nevertheless interested in subtle effects. It is not uncommon for phase to be a concern for fidelity. Still, it is much more common for phase to be ignored, or worse, sacrificed, in order to maximize some other measurement, such as frequency response.

An illustration depicts an orchestra with a two-dimensional
array of 20 microphones wired to a corresponding array of 20 speakers.
An alternate history of what stereo could have been

As musicians, we may or may not care about space at all. But if we do care, we should recognize that space is not something that happens in stereo sound, quadraphonic sound, or even ambiphonic sound. Space is more than these things, and every way that sound exists in space cannot be reproduced by these systems. Conversely, a set of speakers does not have to produce sounds according to the stereo model, and a room does not have to respond to speakers according to the stereo model. Often the most interesting spatializations occur when this model is abandoned and speaker placement and room effects are driven solely by the overall artistic intent.

The Audio Range

The range of our hearing is limited in frequency. The commonly given range of 20Hz–20kHz is a surprisingly good approximation for human hearing, but it doesn’t really reflect how the limits to our hearing work. Our ears detect sounds at loudnesses which change with frequency. Somewhere above 12kHz, this threshold starts to rise, and somewhere around 17kHz, this threshold starts to rise very steeply, then again more gradually after 20kHz. Some studies have detected hearing at very loud levels above 28kHz. But these upper thresholds seem to vary pretty wildly between different people, with some able to hear 20kHz at a loud whisper, and others unable to hear it even at piercing volumes. A similar thing can be said about the lower end of the audio range. Here, perhaps, it makes more sense to extend the commonly accepted range down to 16Hz or lower, as sound tends to have more energy concentrated in the low end.

Note: I am not advising you to create loud sounds outside of the range of hearing. Please be very, very cautious with loud sounds that you cannot hear. They may pass the threshold of hearing damage before they pass the threshold of hearing. The same exposure levels as for the 20Hz–20kHz range should apply.

Despite the fact that these limits to hearing are fairly well documented, the reason for these limits is not completely understood. Bone conduction, where the inner ear is stimulated by vibrations in the skull, rather than the air, seems to be able to transmit frequencies up to 120kHz or even 200kHz. Because of this, some research has theorized that our cilia (the elements of the ear which transform vibrations to electronic impulses) have a very high frequency response, but that our ear canal acoustically filters high frequency sound. Others, however, theorize that high frequencies vibrating through our bones become lower frequencies through processes of resonance or distortion. In musical performances with differing media (water rather than air) or different transducers (bone conduction), frequency limits will be different. While some studies on this exist, outside of air and speakers, even the range of hearing is not yet settled.

But the big problem with 20Hz–20kHz isn’t the specific range or hearing thresholds, but the idea that because we only hear certain frequencies, we can safely ignore what happens with the other frequencies, with no effect to “sound.” This is likely to cause problems. High frequencies might be immediately filtered out by whatever the next stage is, in which case they have no effect. However, high frequencies will have no effect not because they wouldn’t affect what we hear, but because, due to an intentional or accidental feature of the electronics, they are in fact removed without us thinking about it. Otherwise, frequencies higher and lower than we can hear can affect our hearing a number of different ways.

First, when high frequencies excite a speaker, they will contribute to audible distortion. They may do this in other electronic systems as well. This is known as intermodulation distortion, or IMD. In theory, with a completely linear system, high frequencies and low frequencies do not affect each other. But in practice, these frequencies ride together through systems with limits and slight curves, especially the motion of the speaker itself, and different ranges of frequencies have slight effects on each other. Since distortion is always present to some degree in all systems, a high frequency contribution is not necessarily a bad thing, so long as these high frequencies correspond to musical content. Of course, deliberately producing audible distortion by running high frequencies against limits and curves is also possible.

Second, in musical uses, frequencies do not necessarily stay where they originate. A sampler, for example, might pitch a sound recorded at 2kHz to 20Hz instead. In this case, the high frequency harmonic content at the upper 20Hz limit will be the original frequency content at 2MHz, shifting the entire AM radio band into the audible spectrum. More likely, a frequency this high does not exist in the 2kHz recording, and the sampler would sound dark when pushed to these limits. But other mechanisms for shifting frequency ranges, such as modulation or frequency shifting, might preserve frequencies this high.

Third, as noted in the post on Spectra, the decomposition of a sound into individual frequencies is an analytic tool, not a fact about the sound itself. In the lower frequencies in particular, rhythm and frequency are two different ways of looking at the same thing; phase and timing are two different ways of looking at the same thing; and the phase differences created through filtering these low frequencies can affect the clarity ond definition of a given rhythm. 10Hz, for example, is the same as 16th notes played at 150bpm.

The reality is that, although we may be limited by our bodies and our tools, there is no freqency that is wholly irrelevant. All frequencies are a part of sound, without exception, and we can create ways to use them in our music.

Snake Oil, and Other Stories about Hearing

A label with the
words Clark Stanley's Snake Oil Liniment
The label for snake oil

In the late nineteenth century, Clark Stanley began travelling around the United States selling rattlesnake oil as a cure-all. He made numerous claims of its effectiveness, mostly focusing on pain and arthritic relief. In 1893, he presented his product at the Chicago World’s fair, and shortly after went into mass production. Unfortunately for Stanley, in 1917 his product was found to contain no actual snake oil, leading to widespread outrage. Stanley was charged a modest fine, and the whole affair was a huge publicity win for the recently established food and drug administration (FDA), now able to protect the credulous public from being duped by unvetted medical claims.

Snake oil has become a parable for the suggestibility of the public, the untrustworthiness of subjective impressions, and the lies and chaos that wait just beyond the scientific establishment. However, when we take a closer look, the story is not so simple. While Stanley clearly fabricated a complicated lie to sell a product, while people believed him, and while the product as stated had no real research showing its effectiveness, this does not necessarily indicate that his actual product was ineffective. According to the 1917 lab analysis, Stanley’s “snake oil” actually contained mineral oil and tallow along with red pepper, turpentine, and camphor. Camphor and capsaicin, the key compound in cayenne, are now known to relieve pain and inflammation, and are ingredients in a number of FDA approved over the counter pain relief medications. Turpentine is a counterirritant common in liniments at the time. In short, Stanley’s snake oil contained a very plausible list of ingredients for an effective liniment. Snake oil wasn’t made of snake oil, but the public’s perceptions of its efficacy is likely to have been mostly correct. The public was only duped about the reasons, or the models, behind these perceptions.

Nevertheless, the idea has persisted that our perceptions themselves can be fabricated by con men, almost without limit. This poses a problem for discussions about fidelity, as the entire model of audio fidelity is based around perceived sound, rather than objective sound. One solution is to create models of human hearing, and then abstractly apply these models in order to “objectively” determine what someone would perceive. This is often done. However, a cursory look at actual psychoacoustic research will show that generally these models come from first creating a sound and then simply asking people directed questions about what they hear. We don’t have magical models of subjective experience that are somehow untainted by subjective experience.

At some point, even the most skeptical member of Camp Skully needs to explore what happens when a human hears some particular sound system. In a misunderstood analogy with medicine, Camp Skully has devised a listening ritual which tries to eliminate the false snake oil perceptions to which we are supposedly prone. In this ritual, participants gather together and listen to complete systems without actually knowing what systems they are listening to. Then they choose which system they prefer, and in a big reveal that surprises no one, it turns out that money sort of corresponds to quality but also sort of it doesn’t. In a different version, participants just try to reliably identify one system over another. If they can’t do so, then their perception that one system is better is regarded as an illusion not based on hearing. To be fair, many participants in these rituals don’t take them that seriously. But some do.

While I haven’t seen a statistical study, I suspect relatively few members of the fidelity community have participated in anything like this. Instead, they receive this ritual as a myth—in such and such a comparison no one was able to tell the difference. This ritual and myth together form the narrative basis for the difference that doesn’t make a difference, the difference that we can ignore in order to preserve the concept of fidelity. The idea of snake oil, then, usually functions as an attempt to force social cohesion through the (often politely implicit) shame and ridicule of supposedly being a con man or a rube.

This ritual contains a model of perception, separate from psychoacoustic (or even psychological) research. The fidelity listener is exceptionally astute, able to take in the sound as a whole, remember it as a whole, and then accurately, mentally compare it a few moments later. For the fidelity listener, perception is total and passive, and there is no distance between what we can hear and what we can identify. For this listener, active attention is not a part of hearing, or at least attention can be retroactively applied in memory alone. On the one hand, this listener is infinitely suggestible, hearing anything snake oil salesmen tell them to hear. On the other hand, this listener is the perfect Cartesian subject, with a direct and reliable connection between perception and knowledge.

The proponents of this ritual have intentionally designed it to be stringent, to throw out all but the best information. However, the stringency of the test isn’t actually biased towards an inconclusive result, but towards a negative result, towards the difference that doesn’t make a difference. Consequently, this ritual gives us the idea of a relatively incapable listener, and a relatively good system of sonic reproduction. One might contemplate the way that this bias helps the purveyors of hi-fi equipment sell to Camp Skully.

In contrast, in addition to the limits of our hearing, actual psychoacoustic research continually discovers new ways in which human hearing is incredibly good. Certainly our hearing itself is a whole lot better than our capacity to recall and consciously analyze what we hear. Conversely, while there are some parts of audio reproduction systems that might be approaching the threshold of imperceptible differences—well-designed line-level preamps, for example—our systems as a whole are not even close to this ideal, even if we adopt the limiting notion of a perfect listening room and a completely passive listener.

Recordings and Dreams

Everything in our world is constrained by the arrow of time and the separation of space. Something about sound, in particular, seems to accentuate our sense of the ephemerality of experience. But recording promises us something else, and to a certain extent, it actually delivers on this promise. Recorded sound means that otherwise ephemeral experiences can be repeated, over decades and continents. At this point in time, it is almost unthinkable to make music without being immersed in a history of recording over a hundred years long, filled with dead voices, distant cultures, and passed moments. But what can we really replicate? And what disappears and is lost forever?

These are the stakes of fidelity. Knowing we will fail, the proponents of fidelity nevertheless insist that we should have it all, that we should lose nothing, forever. And why shouldn’t they demand this? To outsiders, those who seek fidelity often seem weirdly fussy. There they are, sitting with $10,000 amplifiers cobbled together from tubes, carefully paired with speakers the size of an apartment, tuning their rooms and talking constantly about sound, not music. Surely, they must be satisfied? But instead, they trade in their equipment for things double the cost. They start messing with things that seem tangential and inconsequential. They buy silver cables and build granite platforms. Fussy, at best. But to these audiophiles, these systems are not a means to observe the form of music; they are an incantation to conjure the presence of sound. They want sound to be here, really here—not a shoddy reproduction, but a resurrection.

To whoever is not obsessed, obsession seems like a good way to forget about what really matters. To whoever is obsessed, obsession is the only thing preventing us from losing something forever. There is not a way to stand above this and decide who is right. As musicians, we are free to take on audiophile obsessions, or to leave them aside, either way—we certainly have enough obsessions of our own. But we should learn from these concerns. To some people, sound is going to be sacred. To some people, sound isn’t just a thing we hear, but a future we desire.

References

The opening Hesse quote is from a brand new translation of Steppenwolf from Kurt Beals.

The X-Files captured a particular moment (1990s) of our culture’s thoughts about science and its limits (specifically in The States, but not so different in many other places). The analogy here works best with the first season, which emphasized the inevitable role of desire in shaping all belief, without dismissing the importance of evidence. In subsequent seasons, the first season’s motto “I want to believe” was replaced with “the truth is out there” and the focus slowly shifted to a gradual and unending unraveling of conspiracies and anomolies that was much friendlier to the multiseason TV format.

There are of course multiple differences between analog and digital technologies. The biggest difference one is likely to encounter in synthesis is whether the frequency domain is circular (there is aliasing) or linear (there is infinity). I’ll look at this a little in part 3.

The notion of forms suffuses the entire philosophy of Plato and Aristotle, not just one part. To get a better idea of what this is, first read the Wikipedia page on the theory of forms. Then if you want more probably the best thing to read would be the Phaedo or maybe the Republic (or if you want the weirdest version, the Timaeus). For Aristotle, it’s not a bad idea to start with the same thing medieval philosophers used, Porphyry’s Isagoge i.e. Introduction, followed by the Categories and maybe the Physics. In modern philosophy, some touchstones on similarity, identity, and difference are Fichte, Hegel, Frege, or more recently, Saul Kripke. The way that identity plays out in mechanical reproduction is often touched on with discussions of simulacra—copies without an original. Here you might reference Baudrillard’s Simulacra and Simulation or Benjamin’s “The Work of Art in the Age of Mechanical Reproduction.”

Physics doesn’t say much about similarity and identity, but what physics does say ends up being just as weird as Plato or Aristotle. See, for example, the one-electron universe theory.

The idea of a difference that doesn’t make a difference is modeled on a remark by Gregory Bateson that “what we mean by information—the elementary unit of information—is a difference which makes a difference” See “Form, Substance and Difference,” accessible free in a few places through a web search. Clearly, one could write an entire treatise on the relationship between perception, information, and difference, but here I only mean to be a little evocative of what that treatise might look like.

The passive consumption of media is heavily critiqued in Guy Debord’s The Society of the Spectacle, but Debord focuses primarily on visual media.

For recent work on the localization of very low frequency sources, see Mădălina Anamaria Năstasă, Auditory localisation of low-frequency sound sources, a thesis from Aalto university (2022).

Alan Blumlein’s 1931 patent on stereo sound makes interesting reading. There are a number of techniques he suggests that are still not widely known or used. See UK Patent no. 394325.

For more information on phase coherence and its relationship to the perception of proximity, see David Griesinger, “Phase Coherence as a Measure of Acoustic Quality” on akutek.info (2010). Griesinger is the engineer most responsible for the Lexicon 224 reverb.

Spatial audio is a topic unto itself, and I might address it at some point. I am most interested in non-traditional approaches (vs stereo, quad, 5.1, etc.). For more on this, see the manual for the Quad Mid Side. Michael Gerzon has also done some interesting work that I’m only just now starting to digest.

For more on the upper extended limits of hearing, see Kaoru Ashihara “Hearing thresholds for pure tones above 16kHz” in the Journal for the Acoustical Society of America (2007).

For more on the lower extended limits of hearing, see Jörg H. Mühlhans, “Low frequency and infrasound: A critical review of the myths, misbeliefs and their relevance to music perception research” in Musicae Scientiae (2017).

For more on the limits of human hearing in bone conduction, see Tadashi Nishimura et al, Perception Mechanism of Bone-Conducted Ultrasound and Its Clinical Use in Audiology Research (2021).

The 200kHz figure for underwater hearing comes from the abstract for “Human underwater and bone conduction hearing in the sonic and ultrasonic range” by Michael Qin et al. in the Journal for the Acoustical Society of America (2011). The paper itself seems to be unavailable.

Speculation about the potential effectiveness of snake oil can be found here and there on the web, mostly in special-interest journalism, for example Cynthia Graber Snake Oil Salesmen Were on to Something in Scientific American (2007). I have been so far unsuccessful at tracing down a source in history or anthropology of medicine, if there is one. The journalistic articles are nevertheless convincing, and cite actual medical research. The best credentialed source I have seen discussing this is Robert W. Baloh, Initial Management of Sciatica in Sciatica and Chronic Pain (2018).

  • Contents