Jump to content
  • Dithering Demystified

    By Anderton |

    It's a dirty job to go from high-res audio to 44/16, but someone's got to do it

     

    by Craig Anderton

     

    The ultimate form of digital audio used to have a 16-bit word length and 44.1 kHz sampling rate. Early systems even did their internal processing at 16/44.1, which was a problem—every time you did an operation (such as change levels, or apply EQ), the result was always rounded off to 16 bits. If you did enough operations, these roundoff errors would accumulate, creating a sort of "fuzziness" in the sound.

    The next step forward was increasing the internal resolution of digital audio systems. If a mathematical operation created an "overflow" result that required more than 16 bits, no problem: 24, 32, 64, and even 128-bit internal processing became commonplace (Fig. 1). As long as the audio stayed within the system, running out of resolution wasn't an issue.

    5318ee6696e9b.png.f9ee07c6f825a57fef699cb08ab61ac5.png

    Fig. 1: Cakewalk Sonar allows choosing 64-bit resolution for the audio engine.

     

    These days, your hard disk recorder most likely records and plays back at 24, 32, or 64 bits, and the rest of your gear (digital mixer, digital synth, etc.) probably has fairly high internal resolution as well. But currently, although there are some high-resolution audio formats, your mix usually ends up either online in MP3, AAC, or FLAC format, or in what's still the world's most common physical delivery medium: a 16-bit, 44.1kHz CD.

    What happens to those "extra" bits? Before the advent of dithering, they were simply discarded (just imagine how those poor bits felt, especially after being called the "least significant bits" all their lives). This meant that, for example, decay tails below the 16-bit limit just stopped abruptly. Maybe you've heard a "buzzing" sort of sound at the end of a fade out or reverb tail; that's the sound of extra bits being ruthlessly "downsized."

     

    DITHERING TO THE RESCUE

    Dithering is a concept that, in its most basic form, adds noise to the very lower-level signals, thus using the data in those least significant bits to influence the sound of the more significant bits. It's almost as if, even though the least significant bits are gone, their spirit lives on in the sound of the recording.

    Cutting off bits is called truncation, and some proponents of dithering believe that dithering somehow sidesteps the truncation process. But that's a misconception. Dithered or not, when a 24-bit signal ends up on a 16-bit CD, eight bits are truncated and never heard from again. Nonetheless, there's a difference between flat-out truncation and truncation with dithering.


    SOME AUDIO EXAMPLES

    Let's listen to the difference between a dithered and non-dithered piece of audio. To obtain these examples, I normalized a snippet of a Beethoven symphony down to an extremely low level using a 16-bit audio engine (not 32-bit floating point or something else that would preserve the fidelity, even at low levels) so the effect of dithering vs. non-dithering would be obvious. I then applied dithering to one of the examples, then normalized both of them back up to an audible level.

    Dither Beethoven.mp3 is the file without dithering.

    Dither Gaussian Beet.mp3 is the same file, but with dithering added. Yes, you'll hear a lot of noise, but note how the audio sounds dramatically better.

     

    THE TROUBLE WITH TRUNCATION

    The reason why you hear a buzzing at the end of fades with truncated signals is that the least significant bit, which tries to follow the audio signal, switches back and forth between 0 and 1. This buzzing is called quantization noise, because the noise occurs during the process of quantizing the audio into discrete steps. In a 24-bit recording, the lower 8 bits beyond 16 bits account for 256 different possible levels  between the "on" and "off" condition; but once the recording has been truncated, the resolution is no longer there to reproduce those changes.

    Bear in mind, though, that these are very low-level signals. For that punk rock-industrial-dance mix where all the meters are in the red, you probably don't need even 16 bits of resolution. But when you're trying to record the ambient reverb tail of an acoustic space, you need good low-level resolution.

     

    HOW DITHERING WORKS

    Let's assume a 24-bit recorded signal so we can work with a practical example. The dithering process adds random noise to the lowest eight bits of the 24-bit signal. This noise is different for the two channels in order not to degrade stereo separation.

    It may seem odd that adding noise can improve the sound, but one analogy is the bias signal used in analog tape. Analog tape is linear (distortionless) only over a very narrow range. We all know that distortion occurs if you hit tape too hard, but signals below a certain level can also sound horribly distorted. The bias signal adds a constant supersonic signal (so we don't hear it) whose level sits at the lower threshold of the linear region. Any low-level signals get added to the bias signal, which boosts them into the linear region, where they can be heard without distortion.

    Adding noise to the lower eight bits increases their amplitude and pushes some of the information contained in those bits into the higher bits. Therefore, the lowest part of the dynamic range no longer correlates directly to the original signal, but to a combination of the noise source and information present in the lowest eight bits. This reduces the quantization noise, providing in its place a smoother type of hiss modulated by the lower-level information. The most obvious audible benefit is that fades become smoother and more realistic, but there's also more sonic detail.

    Although adding noise may seem like a bad idea, psycho-acoustics is on our side. Because any noise added by the dithering process has a constant level and frequency content, our ears have an easy time picking out the content (signal) from the noise. We've lived with noise long enough that a little bit hanging around at 90dB or so is tolerable, particularly if it allows us to hear a subjectively extended dynamic range.

    However, there are different types of dithering noise, which exhibit varying degrees of audibility. The dither may be wideband, thus trading off the lowest possible distortion for slightly higher perceived noise. A narrower band of noise will sound quieter, but lets some extremely low-level distortion remain.

     

    SHAPE THAT NOISE!

     
    To render dithering even less problematic, noise shaping distributes noise across the spectrum so that the bulk of it lies where the ear is least sensitive (i.e., the higher frequencies). Some noise shaping curves are extremely complex -- they're not just a straight line, but also dip down in regions of maximum sensitivity (typically the midrange). Mastering programs like iZotope Ozone (Fig. 2) and even some DAWs offer multiple "flavors" of dithering.
     
    5318ee669b4fb.png.0f023d5646f5db2c70f81f2a9cff715a.png

    Fig. 2: iZotope's Ozone mastering plug-in has a dithering section with  multiple types of dithering, noise shaping options, the ability to choose bit depths from 8 to 24bits, and a choice of dither amount.

     

    Again, this recalls the analogy of analog tape's bias signal, which is usually around 100kHz to keep it out of the audible range. We can't get away with those kinds of frequencies in a system that samples at 44.1kHz or even 96kHz, but several noise-shaping algorithms push the signal as high as possible, short of hitting the Nyquist frequency (i.e., half the sample rate, which is the highest frequency that can be recorded and played back at a given sample rate).

    Different manufacturers use different noise-shaping algorithms; judging these is a little like wine-tasting. Sometimes you'll have a choice of dithering and noiseshaping algorithms so you can choose the combination that works best for specific types of program material. Not all these algorithms are created equal, nor do they sound equal.

     

    DITHERING RULES

    The First Law of dithering is to dither only when converting a high bit-rate source format to one with lower resolution. Typically, this is from your high-resolution master or mix to the 16-bit, mixed-for-CD format.

    For example, if you are given an already dithered 16-bit file to edit on a high-resolution waveform editor, that 16-bit file already contains dithered data, and the higher-resolution editor should preserve it. When it's time to mix the edited version back down to 16 bits, simply transfer over the existing file without dithering.

    Another possible problem occurs if you give a mastering or duplication facility two dithered 16-bit files that are meant to be crossfaded. Crossfading the dithered sections could lead to artifacts; you're better off crossfading the two, then dithering the combination.

    Also, check any programs you use to see if dithering is enabled by default, or enabled accidentally and saved as a preference. In general, you want to leave dithering off, and enable it only as needed.

    Or consider Steinberg Wavelab, which has an Apogee-designed UV22 plug-in that inserts after the final level control (you always want dithering to be the very last processor in the signal chain, and be fed with a constant signal). Suppose you inserted another plug-in, like the Waves L3 Ultramaximizer (which not only includes dithering but defaults to being enabled when inserted), prior to the UV22. Unless you disable dithering in the L3 Ultramaximizer plug-in (Fig. 3), you'll be "doubling up" on dithering, which you don't want to do.


    5318ee66a264f.png.ba76cbf3d9bc7e4dfdca2eef233d29dc.png

    Fig. 3: If you use Wavelab's internal dithering, make sure that any other master effects plug-ins you add don't have dithering enabled (in this screen shot, the Waves dithering has been turned off).

     

    However, also note that Wavelab lets you assign plug-ins to always be pre- or post-the final control (or both, if you want them available in either slot). Go to the Options menu, and choose Plug-In Organization. Check where you want the plug-ins available (Fig. 4).

     

    5318ee66a3da1.png.6a4e01a6185e275c91b7414298746196.png

    Fig. 4: The Waves IDR, which is a dithering algorithm, should be inserted only post the final level control. However the Maximizer processors, which are used as effects or for final processing and dithering, are checked so they're available in both locations.

     

    The best way to experience the benefits of dithering is to crank up some really low-level audio and compare different dithering and noise-shaping algorithms. If your music has any natural dynamics in it, proper dithering can indeed give a sweeter, smoother sound free of digital quantization distortion when you downsize to 16 bits.

     

    5318ee66a5515.jpg.016dfb513639bb1d8b333f3efad52fb7.jpgCraig Anderton is Executive Editor of Electronic Musician magazine. He has played on, mixed, or produced over 20 major label releases (as well as mastered over a hundred tracks for various musicians), and written over a thousand articles for magazines like Guitar Player, Keyboard, Sound on Sound (UK), and Sound + Recording (Germany). He has also lectured on technology and the arts in 38 states, 10 countries, and three languages.




    User Feedback

    Recommended Comments

    There are no comments to display.


×
×
  • Create New...