Thoughts about Music Reproduction

If you have not experienced REALLY good music reproduction on a high-end system, I urge you to go to a specialty store (hopefully an independent dealer, but they are sometimes not conveniently located). Usually, they will have a good selection of music, but you can bring your own CDs or LPs if you have favorites. Ask questions about different speakers, amplifiers, and reproduction devices. And listen. If you do this, you MAY get hooked. At that point, you can decide if you want to invest in off-the-shelf components or go down the DIY path.

Of course, a lot depends on your musical tastes. If you lean toward heavy-bass hip-hop or electronica, you can take some shortcuts. But if you are into well-recorded acoustic music, or even classic rock, folk, or jazz, you will probably be a bit more discerning about what comes out of your transducers. I think this is especially true for musicians.

Here is my list of the critical components in an audio chain (in order):

Amplification

You may notice I listed amplifiers last - that is because they are perhaps the easiest to get "right". But they can muck up your listening experience if they are not right. The amplifier should disappear sonically, but this is not so easy to achieve. Amplifiers have to be designed to work with a wide variety of speaker loads, so designers use a one size fits all approach, which can result in compromises. Also, manufacturers love bragging rights and will tout huge power numbers and astonishingly low distortion figures. This is usually marketing hype and may not reflect what an amp actually sounds like playing music.

Of course, there has been a long-running debate among the Objectivists (those who, at one extreme, only believe in measurements) and Subjectivists (those at the other extreme who will say that listening is the only important criteria). Measurements are great and can lead to amplifier designs with vanishingly small amounts of that which can be measured. A pure objectivist will claim that is all that matters and will make the case with data. The subjectivists will say there are things that can be heard but which cannot be measured. The problem that subjectivists have is describing this with terms similar to what wine enthusiasts use to describe different tastes in wine. Language is completely inadequate when it comes to describing differences in taste, smell, or sound, not to mention individual differences in perception. Audio marketing is a combination of the two but reviewers tend toward the subjective and things start to get silly. What is important is how an amplifier sounds with your speakers and your music, and how close you think it is to live performance, free of any artifacts that did not exist in the original.

You may be familiar with the different "classes" of amplifiers: A, B, A/B, D, G, H, variants thereof, and so on. They all have different characteristics and it is not possible to say which is best for you. In general, class A amps are often regarded as the most "musical" but least efficient (meaning they dissipate a lot of energy even when doing nothing). They are typically low-power output devices. To say that class A amps are the most musical is a broad generalization, of course, since it is entirely possible to make a horrible-sounding class A amp that is easily outdone by any of the others. Personally, I prefer class A (assuming it is well done) where it makes the most difference - in the mid-to-high end of the audio range with efficient speakers (which is why I built this amp). For the low end, like woofers, I think A/B is fine (which is why I built this amp) and for subwoofers, I think a class D is just fine. A carefully designed class A amp can sound glorious, whether built with tubes or solid-state devices. They can be either single-ended (meaning a single output device) or push-pull. Class A amps completely avoid what is called cross-over distortion, common in class B amps. That occurs when one output device shuts off before the complementary device can begin to conduct during the output waveform. In the early days of solid-state amps, this was a common flaw and lead to the label of "transistor sound" which was bad. Real bad.

See my article on Tubes vs Solid State for thoughts about the two technologies.

I have lately become interested in Class D for things other than subwoofers, so I might start looking into building a DIY amp using Class D modules if time and budget allow. See my musings on class D in a separate section.

See this article at audioholics.com for a succinct summary of audio amplifier classes..

Distortion

So what about distortion and measurements? There are various types, the most common of which is Total Harmonic Distortion, or THD. Almost all amplifiers will have THD specified. For power amps, the spec is usually for a given output power at its maximum rated level. THD is measured using a single pure sine wave typically at 1 kHz. Hypothetically, if the amp reproduces the sine wave perfectly, it has zero THD. Practically, all amplifiers will exhibit some THD, even if the number is small. Noise measurements are often combined with THD to produce a THD + Noise figure (these days referred to as Signal-to-Noise And Distortion or SINAD), expressed as a percent or decibel level. THD is a gross measure, since it represents the sum of all harmonics present in the output, regardless of their numerical relationship to the fundamental. Generally, even harmonics are less objectionable than odd harmonics, so THD is a bit disconnected from what is pleasing to the ear and can be misleading. 

Frequency Domain vs. Time Domain

Why is the distinction important? When viewing signals on a conventional oscilloscope, we are doing analysis in the TIME domain. One problem with this is that it is impossible to visually determine small (but audible) amounts of distortion (an exception to this is when viewing a system's response to square waves or impulses - see below). You either need to have a distortion analyzer handy or view the FREQUENCY domain of this signal. Usually, this is done by using a Fast Fourier Transform (FFT) on the signal. Our ears are incredible devices, but guess what - they function in the frequency domain by analyzing bands of frequencies. The cochlea is a tiny snail-shaped organ inside our inner ears, filled with fluid and an array of hairs of varying lengths called stereocilia. The hairs vibrate in response to specific frequencies transmitted into the cochlea by the ossicles (three tiny bones) connected to the eardrum. Each tiny hair excites neurons, which then transmit the information to the brain for further analysis.

The cochlea "unrolled" showing how the stereocelia react to different frequencies. From openstax.org

An FFT from DIY Audio contributor XRK971 - actual amplifier under test - see this post

click for larger image

As mentioned, in order to make THD more meaningful, you really need to see a Fast Fourier Transform (FFT) of the harmonic spectrum to see the relative strengths of the harmonic components. Higher-order harmonics (whether even or odd) are more objectionable than lower-order ones, and it has been shown that they should be weighted accordingly. Regardless, THD is at best a measure of those components of a signal harmonically related to a fundamental sine wave, so it should be considered a bit of an artificial characteristic since music never consists of a single sine wave! Here is an example FFT spectrum of an amplifier. The fundamental is the large spike and the smaller ones are harmonics produced by the amplifier. A very low distortion signal source such as I have described here will have few or no harmonics. Also, see this project for how to do DIY distortion measurements on a budget.

Other forms of measurable distortion have been identified over the years. One of the earliest to be measured was Intermodulation Distortion or IM. This type of distortion measurement was an attempt (in the audio realm, at least) to be a more realistic way of characterizing non-linearity in electronic devices. The test consists of two tones (and sometimes 3 tones) fed into a device. If the device misbehaves electrically, it will produce additional signals that ARE NOT harmonically related to either of the two tones and therefore will be annoying to the ear. The sound produced by a device with large amounts of IM will be dissonant and harsh. Usually, there will be a correlation between THD and IM levels, although there are exceptions. There are several types of standardized IM tests, but the most common are SMPTE, typically using 60Hz and 7 kHz tones, and CCIF, typically using 19 kHz and 20 kHz tones. Other frequencies can be used but this adds confusion to the standards and makes it hard to compare apples to apples. Here is an example of an SMPTE test on an older Yamaha integrated amplifier, which has very good THD numbers (.008% at full power) but shows considerable (audible) spurious frequency components at lower power levels due to IM distortion.

IM Spectra of an integrated amp - analysis by Audio Science Review contributor PMA at this link

Note the 'hash" (sidebands) around 7 kHz and 14 kHz which is only about 80dB below the 7 kHz test tone.

Click for larger version.

Square wave testing is a time-domain technique where an oscilloscope can readily show an amplifier's inability to respond correctly to step functions. A "perfect" square wave is fed into the amplifier and the output is observed under various conditions such as frequency, power output, and load conditions. The graphic here shows some extreme conditions. Usually, an amplifier will exhibit a slower rise time than the input due to limited bandwidth. Another common condition is ringing, where the amp briefly oscillates and therefore produces dissonant sound. Manufacturers rarely publish squarewave test results, although many reviewers will do so. For an excellent overview of square wave testing see Rod Elliot's post here.

He includes a DIY schematic for a good squarewave generator that can be built cheaply.

Transient Intermodulation Distortion, or TIM, was identified in the early 1970s as being caused by amplifiers with low slew rates (also mentioned in the literature as Slew Induced Distortion or SID); that is, a lag in the time it takes for an amp to ramp up to the full value of a step-function signal. This was prominent in early small-signal op-amps when they were first used in audio with not great audible results, even though 1 kHz THD measurements we fine. Tube amps were less prone to the phenomenon since tubes are inherently high bandwidth devices. This spec is rarely mentioned by manufacturers since it has largely become a moot point for well-designed amps. The key to minimizing TIM is to design the amp with a bandwidth of at least 100 kHz. This is fairly easily done in class A, B, and A/B amps, but may not be so easy with class D, since they have heavy ultrasonic filtering. Excessive negative feedback, used to correct badly non-linear circuits, is also thought to contribute to TIM (the key, obviously is to design circuits with open-loop behavior as linear as possible, then apply small amounts of negative feedback). Unfortunately, no standards exist for measuring this type of distortion, even though it has been shown to have an audible effect.

There is a TIM "torture test" which was documented by Tektronix many years ago. This consists of feeding an amp with both a 500 Hz square wave and a 6 kHz sine wave. If the amp exhibits bad TIM, lots of undesirable harmonics will show up because as the amp attempts to slew the rising and falling edges of the square wave, the sine wave gets swamped out and therefore not accurately reproduced.  The graphic here shows this effect. There are variations of this test floating around using different frequencies. Again, there are no standards for this. 

For a really good overview of distortion testing, especially for the DIYer on a budget, see Archimago's excellent blog post here. He has a thorough explanation of the types of tests and the implications of the results.

Other Amplifier Misbehaviours

Musicality of Amplifiers

There is actually a technical explanation for some class A amps potentially sounding more "musical". This has to do with the fact that a well-executed class A design will emphasize the second harmonic distortion of fundamental frequencies over odd-order harmonics. Second (and other even-order harmonics) are more pleasing musically. Consider a vibrating string in a musical instrument. The fundamental tone is the frequency at which the string vibrates. But the string will also have overtones, which are harmonics (multiples of) the fundamental. These overtones, along with the construction of the rest of the instrument, are what give the instrument its timbre. Even harmonics predominate in stringed instruments and a small amount of additional even harmonics added by an amplifier are almost impossible to hear. They result in only a very slight perceived change in the timbre of the instrument. Odd-order harmonics, on the other hand, can create a sort of dissonance that could make the sound harsh or unnaturally bright. Another subtlety is the phase of the second harmonic relative to the fundamental created by the amplifier. Negative phase second harmonics are often thought to be more pleasing to listeners. The design of the amp and the characteristics of its components can influence the harmonic phase. Nelson Pass wrote this article which discusses the importance.

I am probably not explaining the importance and subtlety of this harmonic topic very well, so I will defer to a seminal work by Sir James Jeans, published in 1937(!), titled Science and Music, available from Dover Press. This is from a section titled "Vibrations of Strings and Harmonics", starting on page 86:


"The second harmonic adds clearness and brilliance but nothing else, it being a general principle that the addition of the octave can introduce no difference of timbre or characteristic musical quality. When the second harmonic is of equal strength with the first, it produces much the same effect as adding the octave-coupler on an organ or harmonium or playing in octaves, instead of single notes, on the piano.

The third harmonic again adds a certain amount of brilliance because of its high pitch, but it also introduces a difference of timbre, thickening the tone, and adding to it a certain hollow, throaty or nasal quality, which we may recognise as one of the main ingredients of clarinet tone 

The fourth harmonic, being two octaves above the fundamental, adds yet more brilliance, and perhaps even shrillness, but nothing more, for the reason already explained.

The fifth harmonic, apart from adding yet more brilliance, adds a rich, somewhat horn-like quality to the tone, while the sixth adds a delicate shrillness of nasal quality. As the table on p. 73 shews, all these six harmonics form parts of the common chord of the fundamental note, and so are concordant with this note and with one another. 

The seventh harmonic, however, introduces an element of discord; if the fundamental note is c', its pitch is approximately bb'", which forms a dissonance with c. The same is true of the ninth, eleventh, thirteenth, and all higher odd-numbered harmonics; these add dissonance as well as shrillness to the fundamental tone, and so introduce a roughness or harshness into the composite sound. The resultant quality of tone is often described as "metallic", since a piece of metal, when struck, emits a sound which is rich in discordant high tones."


Interesting! So if we translate this analysis to amplifiers (or any other audio reproduction electronics), we can clearly see that in order to have the highest possible fidelity and musicality, we need to minimize the addition of harmonics, particularly the higher-order ones. Each spurious harmonic introduced by the electronics will add coloration to the sound and this could well explain why critical listeners hear differences in equipment. 


So you can see that these objective measurements of amplifiers are not exactly simple, and that has given rise to even more debate among the subjectivists.

Preamps

Do you need a preamp? Preamps are referred to as line-level components - usually involving signal levels 2 Volts RMS or less. It can be passive (meaning no additional signal amplification) if your source devices produce sufficient voltage to drive your power amp. You will need a way to attenuate (as opposed to boosting) the signal to control the volume. Also, you will want a way to select from several sources. All these can be handled by a preamp, but it does not necessarily need to provide gain. See my project page for my implementation. Some preamps feature simple buffer stages with no gain, while most active preamps have selectable gain amounts. In either case, you have active devices in the signal path, so additional distortions/musical colorations will occur. I think if either of these features exist in the preamp, they should ideally be discreet class-A topologies. However, there are some excellent opamps available today that can rival discreet designs, like the TI OPA1656 which many DIY audiophiles are using to build line-level circuits.

All-in-one or integrated amplifiers combine the power amp with a preamp and usually an input selection mechanism. Some people prefer these since they are (theoretically) matched and more convenient, taking up less space. But they also provide less flexibility in terms of being able to swap out just the amp or preamp.

Receivers are much more common because the industry has had to address the audio-visual (A/V) market; that's where the money is. These include tuners, HDMI video switching, and multi-channel surround sound. Some receivers can be very good at audio, but again there are usually compromises to be made in terms of quality. The really good ones end up costing as much or more as separate components. All A/V receivers have digital signal processing (DSP) capabilities for surround effects. In order to get "pure" audio reproduction, you need to shut off or bypass the DSP circuitry. Most receivers have a stereo mode to allow this. If you want to combine hi-end audio with your A/V setup, you can usually find a reasonably priced receiver and then opt for separates for the main stereo listening experience. That's the way I would go if I were to build a combo A/V and audio listening system.

Music Conversion

Music conversion devices are next up the ladder. To get really good sound out of LP records usually requires a significant investment in a good turntable and tonearm, and only the truly brave have attempted DIY construction. If you own a Bridgeport milling machine and a 12-inch metal lathe, be my guest, and good luck. Otherwise, settle for a good turntable from the used market. The cartridge is critical for good reproduction, and they can be pricey too. LPs have been making something of a comeback lately among audiophiles. If you are interested, you should visit a high-end shop and audition some music on LP. If you like what you hear, then consider adding a turntable to your arsenal. For a really fun look at a DIY audiophile who is taking turntables and tonearms to the ultimate level, watch this Steve Guttenberg YouTube episode where he visits "Joe" in his workshop. Of course, LP sources WILL require a preamp and the above discussion about distortion and noise applies to them as well as power amps.

A custom turntable constructed by J.J. Jimink ca. 2008

A Revox reel-to-reel and a Nakamichi Dragon cassette, both from the 1980s

Tape machines were once pretty popular among audiophiles, and several companies supplied tapes only a few generations removed from the original studio masters. That meant that tape copies were pretty close to the original and sounded better than LPs. Most high-end machines were reel-to-reel, with good tapes recorded at 7.5 or 15 inches per second, requiring large reels. When audio cassettes were introduced, several Japanese manufacturers such as Nakamichi produced high-end decks that came close to reel-to-reel quality. Most tape machines have hit the dustbin of history because it was never possible to completely eliminate tape hiss, wow, and flutter. The most ambitious attempt to overcome hiss was Dolby B and C noise reduction; however many would argue that these processes contaminated the sound.

CDs and CD players are waning in popularity, primarily because of streaming formats. Back in the '80s and '90s audiophiles were going crazy doing mods to existing CD equipment in an attempt to overcome some of the inherent limitations of the 16-bit "Red Book" implementations and bad D/A converters (and filters). Some niche manufacturers like Oppo came to prominence and probably took the CD idea as far as it could go. And we had SACD ("Super Audio CD") and HDCD formats - but few music producers seemed interested in creating media for these. Some DVD and BluRay players actually have decent D/A converters and can sound good, but the focus is on video, so your mileage may vary. Almost all these devices have digital outputs, which bypass the internal electronics.  This allows you to hook up a high-quality external D/A converter. All that said, a well-recorded CD can sound very good (unfortunately, there are many badly recorded CDs). Many audiophiles (including me) have converted their CD collections to digital storage by "ripping". If you do that, be sure to use a lossless format like FLAC and "bit perfect" ripping software (see my useful link list for suggestions).

D/A Conversion and Compression

Streaming has taken us into a whole new realm, and new digital formats have become common. Initially, all we had was MP3 (mainly for portable players like the iPod), a lossy encoding format that audiophiles avoid. Very high bitrate MP3 can sound better, but the stream is never "bit perfect" due to lossy compression. Lossy compression made sense back when storage and bandwidth were very expensive, but these days it does not. There simply is no legitimate reason to do lossy compressed audio anymore. In fact, bit rates and resolution (measured in terms of bits) have been increasing. 24-bit 96 kHz (and higher) encoded audio is now commonly available. Note: Only the very best D/A professional-level converters can even approach 24 bits. The very best consumer DACs run about 21-22 bits of dynamic range. This is around 128 dB. The so-called limit of audibility is about 115 dB (down from 0 dB) as a reference. But even converters that can only achieve 18-20 bits of resolution (even though they are advertised as 24-bit) will likely sound better than 16-bit CD quality. 16-bit audio has a dynamic range of 96 dB and must be noise-shaped using dither of the LSB to achieve greater than that. This is a bit of a bandaid and is unnecessary with greater bit depths. 

There are two considerations here: bit depth and sample rate. For an interesting discussion of the historical reason for de facto industry standards, see this article. It explains why we have sampling rates of 44.1 kHz and its multiples, 88.2 and 176.4 vs. 48 kHz and its multiples, 96 and 192. Since the article is about what people in the industry use (users of Pro Tools recording software) there are some interesting comments. One commenter says he is fine using 44.1 for rock but uses 96 for chamber and classical because it sounds better in preserving the subtleties of the acoustic instruments.

I saw a YouTube video recently "proving" that all you need is 16 bits and 44.1 kHz because the output filter in the DAC used was able to convert an ugly staircase signal into a pretty sine wave as viewed on an oscilloscope. Of course - if the original signal is a perfect sine wave with no harmonic components (zero distortion) it will be perfectly reconstructed. Plus, a "brick wall" filter can convert almost any waveform of a fixed frequency into a perfect sine wave. But that means the original signal was corrupted prior to conversion due to inadequate sampling. Another problem is that people misinterpret the Nyquist-Shannon sampling theorem when it is applied to real-world signals and problems with band limiting the signal before conversion. You are throwing away information by filtering the input signal prior to conversion (which is a prerequisite for the Nyquist theorem to hold), especially transients - which are critical in music.

Similarly, there are pundits who will argue that you will never need more than 16 bits and will cite psychoacoustic studies "proving" that people cannot even perceive the difference between 8 bits and 16 bits. This is bunk. I can tell you that all else being equal, there is a clear difference between 16-bit and greater bit depth and sampling rates above 44.1 kHz. Trained listeners (e.g. professional musicians) have demonstrated the ability to perceive differences in double-blind studies. Now the "all else being equal" caveat is a big one, because if you just take a lousy original recording, perhaps one that has been heavily compressed (little dynamic range) or stored on noisy analog tape and then re-mastered in hi-rez, what have you gained? Probably not a lot. Imagine taking an old 78 RPM record and converting it to 24-bit, 192 kHz format - will it sound better? Nope - but at least the imperfections in the 78 recordings will be very accurately reproduced.

The audio recording industry has largely moved to 24-bit 96 kHz recording as a standard, and all D/A manufacturers. support it. The idea is to throw away as little as possible in the audio chain. So if the audio chain is hi-rez from start to finish, then there is a clear benefit. If a studio has made recordings using an end-to-end hi-rez signal chain, why would you want to decimate that down to 16-bit 44.1 KHz? Of course, there are a lot of "remastered" hi-rez recordings made from poor quality masters and mostly do not sound any better than 16-bit.

Back on the subject of compression, none of this means you can't have lossless compression formats. These don't have as small a file size as MP3, but they still save on storage and bandwidth. Popular formats are FLAC (Free Lossless Audio Codec), ALAC (Apple's format), WMA Lossless (Microsoft's), and Ogg, among others. FLAC has become very popular since it is non-proprietary and can encode various bitrate streams, ranging from 16-bit 44.1 kHz (CD) up to 24-bit 192 kHz. Uncompressed formats such as WAV (Waveform Audio Format - developed by IBM and Microsoft - result in much larger file sizes than FLAC, so there is not much benefit in using it for music listening, except for manipulating sound recording. If you are familiar with DAWs (Digital Audio Workstations) used in the recording industry, they typically can manipulate WAV files easily.

The most common encoding method is multi-bit PCM (Pulse Code Modulation) and is supported by all the above formats.

DSD (Direct Stream Digital) is different. DSD was originally created for SACD, then faded away, and is now back. DSD is a single-bit format that uses a very high-frequency carrier rates (2.8 MHz, 5.6 MHz, and even 11.2 MHz). Unfortunately, it is harder to manipulate in the digital realm than PCM, but many audiophiles claim to like the sound of DSD better than PCM. It is not compatible with FLAC and requires a different decoding technique. Unless you have very high-end D/A converters, the differences are small between DSD and high-resolution PCM. Most of the better converters these days can handle both, so you can choose your music format and decide which sounds best in your system.

In fact, D/A converters seem to be getting better and better at ever-lower price points. Here is one example - I recently got a Topping D10s for around $100 for a second small system and it sounds really good. It handles up to 32 bits, 384 kHz, and DSD. A few years ago a D/A converter this capable would have cost $1,000 or more. It even features a socketed opamp in the I/V output stage so you can do opamp "rolling". (n.b. Since I wrote this, Topping has released another $100 DAC, the E30 II Lite, which appears to have even better specs. This proves my point that DACs are getting to be asymptotically "perfect" for little money.)

Subjective reviewers seem to nitpick small differences between low-cost DACs, but I suspect many of the discrepancies are a result of the types of interconnect used (e.g. USB) and the other components, often PCs. Problems such as USB noise conduction and PC ground loops need to be accounted for and can be eliminated by using a good USB isolator.

Topping D10s - an example of a low-cost high-performance DAC.

Credit: Paweł Zdziarski

Here are diagrams illustrating the conceptual differences between Pulse Code Modulation (PCM) and Direct Stream Digital (DSD). PCM - used in CD encoding and many other formats, represents instantaneous values as a multi-bit binary number, ranging in size from 16 to 24 bits. The waveform is sampled at intervals determined by the sampling rate. DSD on the other hand, converts the instantaneous value of a signal into a pulse-width modulated binary signal of only 1 bit (on or off). The sampling rates are much higher. DSD64 uses a sample rate of 2.8224 MHz (64 times that of CDs) and is the standard adopted for SACD. DSD64 can achieve 120 dB of dynamic range within audible frequency bands.

To make things even more confusing, we have DoP (DSD over PCM). In this case, PCM is used as a carrier for the DSD stream, making it compatible with conventional 24/192 PCM transmission channels like USB (Universal Serial Bus, common on all computers), AES (Audio Engineering Society standard interface), and S/PDIF (Sony/Philips Digital Interface - the consumer version of AES).

And finally, I should mention MQA, Master Quality Authenticated. This is a new format that purports to capture all the nuances of original master recordings and requires special hardware (or software on a computer) to take full advantage. Several services in the list below are beginning to offer MQA-encoded music. MQA has resulted in a lot of controversy, ranging from "snake oil" comments to the best thing to happen to audio. So far, I am leaning toward the snake oil end of opinions - do we really need yet another encoding "standard"? BTW - it is proprietary, so not exactly a standard in my book. Just recently (April 2023), MQA filed for bankruptcy protection, so it will be interesting to see what the future holds. Tidal is the largest provider of MQA-encoded streams, so they may take it over.

Update - July 1st, 2023 - According to this article, Tidal has announced that FLAC is now their preferred encoding method for hi-rez streaming, displacing MQA. So while they apparently will continue to support MQA, there is little doubt it is on the way out.

My suggestion is to not worry about the technical differences too much, and just make or buy the newest, highest resolution D/A converter you can afford. DACs have continually improved recently due to competition among chip vendors like ESS, AKM, and Cirrus. This has allowed DAC manufacturers to play the resolution and THD numbers game cost-effectively.

See my latest music streamer project for an example of a very good D/A converter for DIY that costs little and has excellent performance.

The are several online sources of hi-definition audio these days:

The Music Itself

I won't bother pontificating much on the music itself, since personal tastes vary so much, but it is worth touching on the care that goes into capturing the music in the first place. This usually comes down to the producer (and sometimes the artist), the recording setup used, and the recording engineer(s). State-of-the-art recording studios today use the very best equipment in their signal chains. As mentioned, DAWs keep digital domain audio in high-resolution lossless formats, so little is lost in terms of audio information. Usually, with multiple microphone setups and multi-tracking, all the separate tracks are stored individually. Then it is up to the mastering engineer to pull these into a stereo format. This is where things can go off the rails, since the mastering person may be creating the mix the way he or she thinks best, not necessarily what the artist intended or what an audience would have experienced had the artist performed live. Another issue can arise with the type of speakers used to listen to the master re-mix. Since speakers vary greatly, and most commercial studios cater to the more popular forms of entertainment (less so acoustic music), there can be a mismatch with what an audiophile expects to hear. This is why you hear so much variability in recordings. One thing you can do is find something that you really enjoy listening to and seek out other recordings from the same producer/engineering team. This might require some research, e.g. looking up the discography of a particular person.

I was encouraged recently when I read that Jack Vad, the sound engineer for the San Francisco Symphony, uses Linkwitz-designed speakers for the final master recordings of the symphony. See his quote on this page.

Speakers

Finally, speakers. This could be a whole encyclopedia! I'll start a whole new article soon with my opinions.