If you’ve started down the rabbit hole of audiophile gear, you’ve probably come across folks out there imploring you to buy a digital to analog conversion (DAC) unit. It can be a little startling being told that you don’t have the right equipment, but before you go racing off to figure out how much money you’re going to be blowing: read this article first to know if you actually need one. Chances are good that you’re completely fine without it.
This is a long article where I try to be as complete as possible, so feel free to skip around. I just don’t want anyone to see this and feel like they were misled or I glossed over something important.
Editor’s note: this article was updated March 17, 2020 to update the passage on bitrate to address newer audio compression standards, and their ability to perform better with lower bitrates.
What is a DAC?
A DAC simply converts a digital signal into an analog one so that your headphones can then create sound. It’s that simple! Most DAC chips are found in the sources of whatever you’re listening to, and generally run a manufacturer from $3 to $30. At this point, it’s a very basic component of any smartphone, though the headphone jack seems to be a dying feature (Editors’ note: a pox on your house, Apple).
A DAC simply converts a digital signal into an analog one so that your headphones can then create sound.
Much like headphone amplifiers, standalone DACs came about as a response to poor audio quality at the consumer level. Back in the day, it was a lot tougher to find good hardware, and nerds like me had to deal with devices that couldn’t keep up with higher-end headphones and speakers. Sometimes the DAC assembly would be improperly shielded—introducing staticy noise—or it’d be a little too cheap, making the output kinda crappy. Lower sample rates, badly encoded MP3s… there were tons of things that children of the 80s and 90s had to deal with when it came to audio. Who wants to listen to low-quality tunes?
But digital music has come a long way since then. Better tech has made shortcomings of even the cheapest chips almost nonexistent, while digital music has exploded in quality past the point of diminishing returns. Where it used to be true that your Walkman’s or laptop’s internal DAC chip wouldn’t be suitable for high-bitrate listening, there are plenty of more compact units nowadays that can keep up.
When do I need a DAC?
The reason you’d get a DAC today is that your source—be it your computer, smartphone, or home system—is introducing noise or incapable of outputting sound at the bitrate of your files. That’s it. I know that’s a really anticlimactic summary, but that’s really the long and short of it. The only other time you could possibly want something super high-end is if you’re recording audio for professional applications, but even then the equipment used for processing it can handle it relatively cheaply.
Because DACs are a largely spec-driven item, you can almost always pick out the one you need simply by looking at the packaging. FiiO makes plenty good products for cheap, and if you want an amplifier to go along with the DAC so you never have to worry about either, their E10K is a solid pickup for under $100. You could also decide to throw money at the problem by picking up an ODAC or O2 amp + ODAC combo, but that may be overkill. But seriously, don’t sink too much money into this. It’s just not worth it.
How does a DAC work?
Now that you know the why of DAC, let’s delve into the how.
All audio, whether it’s stored on vinyl or in an MP3 is a compression wave when it’s played back. When computers record an analog signal, typically it will be displayed in what’s called a waveform, or a representation of the wave where the Y axis is amplitude (how powerful the wave is), and the X axis is time. Each wave will have a crest and valley—called a period—and how many periods there are in a second is called frequency (displayed as Hz). If you’ve heard that word before, you know that what frequency a sound is also corresponds to what note it is. The higher the frequency, the higher the note.
The job of the DAC is to take a digitally stored recording and turn it back into an analog signal. To do that, it needs to translate the bits of data from digital files into an analog electrical signal at thousands of set times per second, otherwise known as samples. The unit then outputs a wave that intersects all those points. Now, because DACs aren’t perfect, sometimes this leads to problems. These problems are jitter, aliasing, narrow dynamic range, and limited bitrate.
Before launching into the nuts and bolts of how everything works, you need to know three terms: bitrate, bit depth, and sample rate. Bitrate simply refers to how much data is expressed per second. Sample rate refers to how many samples of data are taken in a second, and bit depth refers to how much data is recorded per sample.
What is jitter?
I’m going to preface this section just like I addressed it in the audio cable myths article: Jitter is mostly a theoretical problem at this point, and extremely unlikely to rear its head in any equipment made in the last ten years. However, it’s still useful to know what it is and when it might be an issue, so let’s dive in.
So remember how I said that sample rate can lead to some problems? Jitter is one that gets a lot of attention, but not much understanding. Jitter is a phenomenon that occurs when the clock, or what tells the DAC when to sample, doesn’t work as well as it needs to. When the sample points aren’t happening when they should, this can lead to a change in pitch for short periods of time. The higher the note that’s being reproduced, the higher the likelihood that this problem might occur.
You don't need to worry about slight imperfections in notes near 20kHz because in all likelihood you can't hear them anyway.
However it should be pointed out that this is another one of those problems that isn’t as common anymore because DAC units of today are so much better than those of the past. Jitter tends to only happen at super-high frequency notes because those notes have the shortest wavelengths. However, what makes high-frequency notes more susceptible to this type of error also makes them less likely to be heard: most people over the age of 20 can’t hear the notes where jitter is most likely to occur.
You don’t need to worry about slight imperfections in notes near 20kHz because in all likelihood you can’t hear them anyway.
What is aliasing?
Basically, sometimes a sound that’s really really high in frequency like a cymbal shimmer, harmonic, or other high note will have this strange warbling or oscillating sound that wasn’t in the original recording. What’s happening is that the DAC is accidentally creating a lower frequency note because the sample rate isn’t high enough.
How do you avoid this problem? Increase the sample rate of course! The more data points you have, the less likely an error will happen in a given set of frequencies. However, there is a point where this simply doesn’t audibly help anymore. Essentially, you can eliminate this problem if you’re able to sample at least twice per period, thereby forcing sampling errors to exist only in the highest frequencies that you’d likely be unable to hear anyway. Considering that the uppermost limits of human hearing range from 12-22kHz (as in, 12,000 to 22,000 periods per second), doubling that rate nets you somewhere within 24-44 thousand samples per second, or 44kHz. That last number sound familiar? It should: 44.1kHz is the most common sample rate for MP3 files!
What is bit depth and dynamic range?
If you’ve listened to really old MP3 files or crappy MIDI music from your old consoles, you’ll probably notice that they can’t really ramp up volume in a given music track all that well, or that competing instruments are really really difficult to pick out if they’re all going at once. This is what bad dynamic range sounds like. Dynamic range in this instance simply refers to the difference between all possible volumes of sounds in a given file.
What governs the theoretical limits of the dynamic range of an audio file is the bit depth. Basically, every single sample (discussed above) contains information, and the more information each sample holds, the more potential output values it has. In layman’s terms, the greater the bit depth, the wider the range of possible loudness of notes there are. A low bit depth either at the recording stage, or in the file itself will necessarily result in low dynamic range, making many sounds incorrectly emphasized (or muted altogether). Because there’s only so many possible loudness values that a sound could have inside a digital file, the lower the bit depth, the crappier the file should sound however you listen to it. So the greater the bit depth, the better, right?
Well, this is where we run into the limits of human perception once again. The most common bit depth is 16, meaning: for every sample, there’s a possible 16 bits of information, or 65,536 integer values. In terms of audio, that’s a dynamic range of 96.33dB. In theory, that means that no sound under 96ish dB should be deleted or incorrectly assigned a loudness value.
While that may not sound terribly impressive, you really need to think hard about how you listen to music. If you’re like me: that comes from headphones 99+% of the time, and you’re going to be listening to your music at a volume much lower than that. For example, I try to limit my sessions to about 75dB so I don’t cook my ears prematurely. At that level, added dynamic range isn’t going to be perceptible, and anyone telling you otherwise is simply wrong. Additionally, your hearing isn’t equally-sensitive across all frequencies either, so your ears are the bottleneck here.
While I'm a super big crank when it comes to silly-ass excesses in audio tech, this is one point I'm forced to concede. However, the necessity of 24-bit files for casual listeners is dramatically overstated.
So why do so many people swear by 24-bit audio when 16-bit is just fine? Because that’s the bit depth where there theoretically shouldn’t be any problems ever for human ears. If you like to listen to recordings that are super quiet (think, orchestral music)—and you need to really crank the volume in order for everything to be heard—you need a lot more dynamic range than you would with an over-produced, too-loud pop song would in order to be heard properly. While you’d never crank your amp to 144dB, 24-bit encoding would allow you to approach that.
Additionally, if you record music, it’s always better to record at a high sample rate, and then downsample, instead of the other way around. That way, you avoid having a high-bitrate file with low-bitrate dynamic range, or worse: added noise. While I’m a super big crank when it comes to silly-ass excesses in audio tech, this is one point I’m forced to concede. However, the necessity of 24-bit files for casual listeners is dramatically overstated.
What’s a good bitrate?
While bit depth is important, what most people are familiar with in terms of bad-sounding audio is either limited bitrate, or aggressive audio compression. Ever listen to music on YouTube, then immediately notice the difference when switching to an iTunes track or high-quality streaming service? You’re hearing a difference in compression quality.
If you’ve made it this far, you’re probably aware that the greater the bit depth is, the more information the DAC has to convert and output at once. This is why bitrate—the speed at which your music data is decoded—is somewhat important. If the bitrate is low, it’s possible that not enough data will be converted to create the analog wave, meaning less information is converted, meaning you hear crappier audio. It’s really as simple as that.
320kbps is perfectly fine for most applications... and truth be told most people can't tell the difference.
So how much is enough? I usually tell people the 320kbps rate is perfectly fine for most applications (assuming you’re listening to 16-bit files). Hell, it’s what Amazon uses for its store, and truth be told most people can’t tell the difference. Some of you out there like FLAC files—and that’s fine for archival purposes—but for mobile listening? Just use a 320kbps MP3 or Opus file; audio compression has improved leaps and bounds in the last 20 years, and newer compression standards are able to do a lot more with a lot less than they used to. A low bitrate isn’t an immediate giveaway that your audio will be bad, but it’s not an encouraging sign.
If you’ve got space to spare, maybe you don’t care as much how big our files are—but smartphones generally don’t all come with 128GB standard… yet. But if you can’t tell the difference between a 320kbps MP3 and a 1400+kbps FLAC, why would you burn 45MB of space when you could get away with 15?
Frequently Asked Questions
Yes, using the DAC in your receiver will give you great audio from your CD transport. Although theoretically the coax and optical connections should be the same, the optical cable is generally considered to be a cleaner connection as it electrically isolates the two components.
On some older phones: sure. But the errors and noise that you're likely to come across in this situation will be minor at best—not really something that you'll hear in the din of a commute, or wherever else you go with your smartphone as your main source of music.
No. Do not buy a DAC for Bluetooth headphones, as they will already have a DAC chip inside to handle converting the digital signal to an analog one to send to the headphone's drivers. A second DAC would be redundant.