A Solution to Wow/Flutter?

I've only used the Apogee AD/DA 16x in our rooms. It was too edgy for me as well. If you're strictly listening to phones, I'd suggest a nice tube 'phone pre, or even the Mytek Stereo 192/DSD DAC. It has a great 'phone output.

You're listening observations were on the money. I'm impressed!
 
Bruce, thanks. It's nice to know I'm neither deaf nor crazy.

The Mytek seems like my best move, especially since it would eliminate the need to convert SACD to PCM. I gather the AudioGate settings i mentioned aren't responsible for the edgy sound, yes? I'd hate to think I'd have to reconvert everything. None, BTW is clipped. I use AG's function that analyzes each cut and sets an average non-clipping gain. Would this have any deleterious effect?

Marshall
 
Great read indeed! A few months ago I went researching Wow and Flutter effect to understand their perceptual effects relative to jitter in digital audio. One great paper I found had a fascinating point related to this device in that it also said the standardized weighting for Wow and Flutter was wrong since it had not kept up with time in the way the new (at that time) tape mechanisms generated a lot high frequency variations. He went through the details of what caused that.

This is important since low frequency variations in tape speed creates sidebands that are masked by the signal and hence, even high levels of it are very much tolerated. It only becomes an issue when the frequency becomes so low that we detect the variation in speed (a different problem than jitter).

All frequency variations create sidebands, that's how FM works... as for the last statement that's a bit upside-down ... in truth the sidebands are NOT masked by the signal since they are not harmonically related. At the least it creates thickening ( "generation loss") and in the case of higher flutter rates (100Hz -4kHz) the sidebands are audible as intermodulation distortion, beat frequencies within the audio.
 
Sample A sounds distinctly different than B, and B seems like what I'm more used to hearing. In fact, it is almost identical in character to an 88k SACD rip I have of the same song. If you play B after A, B starts out sounding a little flat.

Sample A overall, is clearer, more open with more inner detail. It's easier to hear individual voices in the background vocals, individual cymbals and instrument placement. Sample B by comparison (and my 88k version) are slightly veiled and 'scrappy' sounding. In this context 'scrappy' means a little more distorted especially around sibilance-over-other-sounds.

Sample A also demonstrates how different the audio on Stevie's lead vocal track is compared to other tracks.

It's hard to imagine that Wow&Flutter were the only thing removed. It sounds like a generation closer to the master.

--Bill

You are correct, and that's absolutely typical of the commentary and characterization. Yes it's closer to the master. It's our postulate that what was thought of as magnetic loss or electronic loss is actually IM caused by the transport. Happens every time. This particular sample is more subtle than most, and was not captured with our heads or preamp, but rather was a trace remnant in the file provided by the mastering shop.

Flutter is not just a gargle - that's not what we're about. Which by the way the software stuff can't touch without risking drying out the vibrato. We see tremendous amounts of flutter in the range of 100 - 1000 Hz, and often if the bias is strong enough, the scrape flutter at 2.5kHz or so.. all of this is cross-modulating the audio - just as a ringmodulator would do. That's why it's such a surprise on well made masters - we didn't really know what flutter could do, damage-wise until we could remove it. We always knew the tape sounded different from the console, even when the tones were perfect and the rec/repro was maximally flat. The transport is a distortion source.
 
Interesting comparo!

On the processed version when Lani Groves (?) Gloria Barley (?) sings "You are the apple of my eye," I can hear here pull her lips apart at "app-le," almost as if she's taking in a small breath. (good technique to avoid popping the P?)

That's blunted on the unprocessed file. The processed file also seems to un-smear the closed high-hat, making it easier to hear the (very) damped metallic ring of it. The effect is even more apparent on the ride. At about 1:00, the drummer works his way from the middle of it toward the bell. This is much more apparent on the processed file.

The final effect I'm less sure of, possibly due to the Meniere's: The unprocessed file has more of a swishy, phase-y throbbing between L and R in the bass range--a pulsing, alternating kind of pressure. The processing diminishes it. Wonder if that's the sound of the wow going away.

Marshall

Yup. That's true, too. The bell sound of cymbals (the pang vs the swish) is masked by fast flutter and scrape flutter. The bass modulation is slight wow. It's working. Again, this is a rather subtle improvement compared to what's possible, but you guys are hearing exactly the correct attributes.
 
Thanks for listening.
Hoping soon that we can put together a more dramatic test... it's all about the s/n of the FM modulation... Which in this case was pretty bad. We didn't expect much of this test, and I would maybe have wanted you all to hear something a little more obvious, but thanks to those who posted here - the comments are right on the money.

FWIW on a major catalog project we worked on recently the A/D of choice was the Mytek. Good that we can hear too, LOL.
 
Yes, you use your own A-D converter at an optimized sample rate in relation to your bias freq. If your bias freq is 120k, then you will need to record to at least 352.8kHz because 192k is too low (Nyquist).
DSD64fs wouldn't work because of the UHF noise, but DSD128fs might. We'll have to see.

Actually we do output a 24kHz carrier that has all the wow and flutter encoded within it, and that can be recorded at 88.2/24 if need be. To capture the bias "raw" you need the higher fs. Noise shaping in DSD converters moves the noise into the area we are trying to dig out. We could potentially mult the signal such that the bias was recorded PCM, and married later to the DSD audio for dewow/deflutter, but we haven't done it yet. It's just math.
 
Welcome to the WBF forum. Jamie.
All frequency variations create sidebands, that's how FM works... as for the last statement that's a bit upside-down ... in truth the sidebands are NOT masked by the signal since they are not harmonically related.
Masking has no reliance on harmonic relationship. All it cares about is whether the signal is in the shadow of another louder signal or not. Here is an example of the point I was making and how masking works:

i-bSLwHw3-XL.png


The fat gray is our music tone. The red and blue are two distortion sidebands created by speed variations. They are both at identical levels for the sake of this discussion. The blue represents the lowest frequency variation and therefore is closest to our source. We see that it lands within the masking threshold. Now if we change the frequency of speed variation and increase it, we get the red version. That same distortion is now audible because it falls outside of masking area. Worse yet in this instance, it falls in the area where the ear is most sensitive as represented by the threshold of hearing curve. So it has two things going for it as far as audibility.

As I noted, it is a lucky thing that our hearing system works this way. This is one of the main reasons why people have preference for analog systems like LP and tape despite their massively higher measured distortions.

At the least it creates thickening ( "generation loss") and in the case of higher flutter rates (100Hz -4kHz) the sidebands are audible as intermodulation distortion, beat frequencies within the audio.
My note was relative to the common manifestation of speed variation which is in the very low frequencies of a few hertz where its amount can be very large yet surprisingly inaudible as distortion in the way we think about it. To the extent the system has high frequency variations of the rates you mention then yes, it will occur at higher frequencies and masking is not as helpful -- just like in our digital systems. That was the point of my first paragraph you quoted in how tape transports had changed in the later years and introduced high frequency variations.
 
Welcome to the WBF forum. Jamie.

Masking has no reliance on harmonic relationship. All it cares about is whether the signal is in the shadow of another louder signal or not. Here is an example of the point I was making and how masking works:

i-bSLwHw3-XL.png


The fat gray is our music tone. The red and blue are two distortion sidebands created by speed variations. They are both at identical levels for the sake of this discussion. The blue represents the lowest frequency variation and therefore is closest to our source. We see that it lands within the masking threshold. Now if we change the frequency of speed variation and increase it, we get the red version. That same distortion is now audible because it falls outside of masking area. Worse yet in this instance, it falls in the area where the ear is most sensitive as represented by the threshold of hearing curve. So it has two things going for it as far as audibility.

As I noted, it is a lucky thing that our hearing system works this way. This is one of the main reasons why people have preference for analog systems like LP and tape despite their massively higher measured distortions.

My note was relative to the common manifestation of speed variation which is in the very low frequencies of a few hertz where its amount can be very large yet surprisingly inaudible as distortion in the way we think about it. To the extent the system has high frequency variations of the rates you mention then yes, it will occur at higher frequencies and masking is not as helpful -- just like in our digital systems. That was the point of my first paragraph you quoted in how tape transports had changed in the later years and introduced high frequency variations.

Actually that's not how it manifests. Here's the FM theory: there should be two sidebands in your graph around the music tone, + and - beat frequencies. An A440 flute tone with a 0.1% flutter of 60Hz and 120Hz (typical with AC cogging motors) will have sidetones of 500 660 and also 380 and 320 at about -60db. Which is not healthy, nor is it easily masked - the first recording we noticed this on was a piano/clarinet duet and the beats in the clarinet were quite obvious, after hearing them absent... It's a ring modulation.
 
Actually that's not how it manifests. Here's the FM theory: there should be two sidebands in your graph around the music tone, + and - beat frequencies. An A440 flute tone with a 0.1% flutter of 60Hz and 120Hz (typical with AC cogging motors) will have sidetones of 500 660 and also 380 and 320 at about -60db. Which is not healthy, nor is it easily masked - the first recording we noticed this on was a piano/clarinet duet and the beats in the clarinet were quite obvious, after hearing them absent... It's a ring modulation.
I didn't say FM creates a single sideband but rather, expanding on how maksing works in general. As to your point, I wrote: "My note was relative to the common manifestation of speed variation which is in the very low frequencies of a few hertz where its amount can be very large yet surprisingly inaudible as distortion in the way we think about it. To the extent the system has high frequency variations of the rates you mention then yes, it will occur at higher frequencies and masking is not as helpful -- just like in our digital systems. That was the point of my first paragraph you quoted in how tape transports had changed in the later years and introduced high frequency variations."

The examples you cite are clearly more than a few Hertz and agree with my statement and explanation of masking and how it is not as effective there. So please don't post again saying we disagree :).

The reason for my original post was that often analog systems are compared to digital and people wonder why we worry about picoseconds of jitter where far larger amounts of it in analog systems is not talked about as being problematic. Much of the reason for that is masking. I wrote a much longer post explaining the difference on another forum as *it relates to very low frequency wow and flutter*. Here it is:


----------------------------------------------------

OK, some free time to reply to Terry's request on the relationship between Flutter and Jitter.

Let me start this with saying I have not studied the perceptual effects of flutter in analog equipment. It has not been an area of interest for me in the past and likely won't be in the future. While I am hoping to add fair bit of data to this discussion, my goal is not to quantify what flutter sounds like but rather, whether its concepts apply to jitter in digital systems. In that sense, some of this reply is open-ended and I welcome others' feedback on what it means as a complete picture.

With that out of the way, let me frame the question as I understand it. The thread again started with me saying very low frequency (50 to 100 Hz) jitter is not as audible as high frequency ones. Arny objected and went on to cite references regarding flutter instead, and how as its frequencies increase, the opposite is true. Unfortunately his references are to papers not available to general public and hence I have not read them. Regardless, I am going to take his word for it and attempt to analyze the situation.

FM Modulation

Arny has correctly stated that both jitter and flutter are Frequency Modulation or FM for short (if this were an advanced topic, we would talk about jitter also being a form of phase modulation but that distinction is not material here). What is frequency modulation? It simply means you take a signal -- called a carrier (Fc) -- and change its pitch or frequency proportional to another signal (Fm). The terminology makes sense in the context of RF transmissions but not so much in our audio world. To clarify then, our "carrier" in this discussion is our music. The modulator signal (Fm) is flutter or jitter. It is the thing that forces the change in frequency of our music tones.

As an example of FM modulation, if I have a 1 KHz tone and a 2 Hz flutter, the 1 KHz tone changes in frequency every half a second and keeps repeating. You will hear this as "WOW" which is the name of flutter when its frequency is very low. It kind of sounds like wow wow wow.

As an aside since complex sounds like music can be decomposed into a series of sine waves, we have countless carriers, each being modulated by the flutter/jitter. So even though we usually show a single tone as the carrier in demonstrations of jitter effect, our music represents many carriers and hence the real picture is far more complex (and hence has more distortion products). But we digress.

I hope everything is clear so far because from here on, it is going to get complicated :).

Perceptual Effects and Modulation Index

To know the effect of Frequency Modulation, we need to look at the signal in frequency domain. In time domain we simply see the carrier shrinking and expanding with the modulator which is uninteresting as far as deducing distortion products. In frequency domain, an entirely different picture exists where we see new frequencies created that have nothing to do with the original signal. The nature of these new frequencies/distortion products depend on a key parameter called the "modulation index:"

Modulation index = peak frequency deviation / Fm

The denominator is the flutter or jitter bandwidth which in case of simple simulations of a single tone, is equal to the frequency of that tone. So this part is easy to understand.

The numerator is the maximum change in the carrier (or music signal) frequency. What determines this is the *amplitude* of the modulation signal. After all, that is what frequency modulation is. The louder the modulation signal, the more the carrier shifts in frequency. And the fainter it is, the less it does that.

The mathematics tells us that as the modulation index increases, the relevant distortion products multiply in number and relative amplitude (in reality they are always there but the higher order values are so close to zero as to not matter in cases of lower modulation index). In case you are curious, the representation of the distortion products is the Bessel functions. Going on a limb and assuming you don't know what a Bessel function looks like, here is a great visualization which shows what happens as we go from very low modulation (called narrowband FM) to very high (wideband FM):

11.gif


Notice our original tone in the middle as the animation starts (it is above the green letter "v"). We see that our simple tone quickly becomes a complex waveform including having its amplitude reduced and varied as modulation index rises. Indeed, this is the heart of "FM Synthesis" technology used to generate artificial sounds in music synthesizers and PC software alike. It is a simple way to create complex waveforms.

While the dancing graph is pretty it may still be hard to visualize what all of this sounds like. Fortunately I found a nice simulation from Columbia University that shows what happens as the modulation index goes from 6 to 0 (high to low), with a carrier of 100 Hz and a modulator frequency of 280 Hz: http://music.columbia.edu/cmc/MusicA...r4/bell.fm.mp3

As the name indicates, one is able to simulate the sound of a bell with the mere action of changing the amplitude of the modulator! The ear is not hearing the tone change at the beginning but an entirely different sound due to combination of the original tone plus all of the new distortion products. All the while, both the carrier and modulator frequency stay constant. The only thing changing is the amplitude of the modulator. Yet we don't perceive that as an increase in distortion after some point but a wholesale change in sound.

Modulation Index effect on Flutter and Jitter

Recall the mathematics of jitter from Dunn's seminal paper that I have often used to show how we drive the jitter spectrum and its amplitude. Here I have reproduced it with a circle to highlight something important:

i-jTdvZnq-X3.png


He starts with the textbook definition of FM modulation which is to change the frequency of the carrier. He then takes a massive shortcut without which, the solution would have been orders of magnitude harder and result in the aforementioned Bessel functions. That shortcut comes courtesy of considering that jitter amplitude in digital systems is usually very weak relative to the original signal. Recall for example the example I used where jitter was at -80db.

When we do that, we throw out a bunch of terms out of the equation and are left with the solution that there are only two sidebands of interest with their frequency being +- jitter frequency as I explained in my last long post. Dunn then calls this outcome "AM modulation" since that is the result of amplitude modulation. In reality, we still have an FM modulator but our modulation index is so low that all the rest of the sidebands are down to zero and hence of no interest. This is what the dynamic simulation above shows the minute the cycle starts -- the original tone plus two sidebands.

Now it is time to look at flutter. As Arny has correctly put it, the amplitude of flutter in analog world as a modulation signal is quite large relative to jitter assumption above. Here is his positioning:

arnyk said:
The jitter due to wow and flutter in analog equipment at its best, is about 1 million times stronger than the jitter in the worst digital equipment. Another way to describe 1 million times stronger is to say that it is 60 dB greater.

Recall that modulation index is proportional to amplitude of the modulator and inversely proportional to its bandwidth/frequency. Therefore when it comes to flutter our numerator has shot way, way up. Therefore we are no longer living in the simple world that Dunn painted in digital domain. This means the other distortion terms do not go to zero and we have a full blow Bessel situation on our hand, if there is such a term :). The spectrum now has a ton of sidebands as the simulation shows with increasing modulation index.

The effect is magnified in other ways. Assume someone is performing audibility tests for flutter. Naturally they start with very low frequencies such as 0.5 Hz to say 20 Hz. Because the numbers are so small, they (a) make the modulation index bigger since it is in the denominator. But further, as you vary them, their effect on the modulation index is quite a bit bigger. For example, as you go from 1 Hz to 4 Hz, the change in modulation index is 4:1. That large change in modulation index radically modifies the resulting audio spectrum and hence level of audibility.

Such is not the case in digital/jitter. For one, the frequencies of interest are not that low but more importantly, the numerator for modulation index is so much smaller that changing the modulator frequency does not impact it that much. We may go from 0.01 to .02 modulation index whereas flutter may go from 2 to 4. The former does not create new sidebands whereas the latter certainly does. But further the creation of sidebands also changes the amplitude of the carrier itself. So you get hit coming and going as modulation index changes. Again you can see this in the visualization as the center tone changes its amplitude in unison with higher modulation indices.

Frequency Unmasking

As much as I have talked about frequency masking, it is important to understand that there are cases where distortion products become unmasked due to another factor and hence, potentially become audible. Such is the case with modulation index. Look at the two snapshots of our first visualization at the beginning (very low modulation index) and end (high modulation index):
i-tj86DwQ-X3.png


Notice two things. First, the amplitude of the carrier or in our case, the original music tone has gotten attenuated. And second, the bandwidth of the sidebands per earlier description has been hugely widened. So not only the masking shadow around the primary music tone has much less amplitude due to shrunk strength of the carrier but that there are now components easily going past the skirts of masking. No amount of masking will cover the second scenario of higher modulation index.

Here is another visualization on the same theme:
fm-spectrum-02.gif


Look at the modulation index 2 where the carrier is so much less strong than the sidebands. Clearly no masking occurs for the strongest sidebands next to the carrier. Only marginal effect exists for the third sidebands if any.

Depending on the carrier, modulation frequency and severity of modulation index, it is possible to actually fully suppress our original music tone and be left with nothing but distortion products! In that situation, there is no masking whatsoever from the music itself (although the sidebands mask each other by varying degrees).

So let's say our flutter frequency was the same as line frequency at 100 Hz. Unlike jitter, we have more sidebands due to high modulation index coupled with reduced power of our original tone. The sidebands may extend for many multiples of 100 Hz (including some inharmanics). So any masking is going to only cover part of the distortion spectrum and some distortion will lilkely leak through.

Again, none of this at play with jitter because the modulation index is so much lower and hence, we have a full amplitude carrier (music), and only two sidebands of much lower amplitude being strongly masked in the case of low frequency jitter.

Wow and Flutter Perceptual Model

As I noted, my goal is not to convince you of audible artifacts of Wow and Flutter. But let me share some data that I do have since the topic has been so much discussed by Arny.

The first point is that when we perform Wow and Flutter measurements, we use a standardized frequency weighting. Arny noted the same in his later posts (drawing from the Wikipedia article):
arnyk said:
I double checked the reference, and the quote from Villchur and Allison was correct and in context. It is also incorrect in the light of a number of later references including Zwicker and Fastel etc.

iec386.jpg

Since it shows weighting designed to correspond to the sensitivity of the human ear, it is showing an attenuation of low frequency noise. The human ear essentially attenuates its sensivity to FM noise at low frequencies.

The curve of course does NOT agree with what started this entire debate as Arny corrects himself above. He originally said that as the frequencies of flutter increased, audibility kept declining. The CCIR/DIN/IEC curve above clearly articulates a different situation where there is a peak at 4 Hz (i.e. most annoying).

Now let's refer to another slide from the FHG presentation I originally cited which actually builds on Arny's reference book on perceptual models as footnoted:

i-T9VC8Jk-X3.png


As you see in chart (c) for FM modulation, there is a sensitivity peak at 4 Hz which is hypothesized as being the case due to human evolution of trying to understand speech which itself is a manifestation of temporal masking (see the slide I post earlier). This nicely agrees with the industry recommendation of paying more attention to 4Hz as the most critical frequency for wow and flutter.

Small and Big Signal Theory

The Electrical Engineers reading this are familiar with analog circuit design where we have the concept of "small signal" and "large signal" analysis. It would be deadly to take small signal rules and apply them to large signal usage of the circuit. The same is in play here. Just because "FM modulation" is used in both jitter and Flutter, we don't get to blindly apply the rules of one to the other. The amplitude of the modulation index radically changes the nature of the situation and its analysis.

Orthogonality of Jitter and Flutter Distortions

A key consideration here is that jitter has a characteristic that flutter never has: data dependency. A tape recorder wow and flutter does not change based on what is being recorded or played. The variations come from mechanical aspects of the machine. The electrical signal driving the head makes no difference. Not so with jitter. Jitter can be induced because one transmits certain data pattern.

Recall the cable example I gave earlier. Such a modulator is not periodic and can cause distortion products not modeled by the above analysis. And it certainly is not anything predicted from flutter world.

Further, digital systems MUST be bandwidth limited to half of their sampling frequency. Bad things happen when you don't do that. Jitter adds sidebands that are at +-frequency of jitter relative to music spectrum. It is very easy to imagine cases where the addition of jitter to music pushes us past the maximum allowable bandwidth. When you do that in digital systems, the new frequencies fold back to into the audio band. This will have an entirely different effect than in analog system where it either sits in ultrasonic band or gets filtered out.

As an example, an 8 KHz jiter tone added to 21 Khz creates an upper sideband at 29 Khz. In CD with bandwidth of 22 KHz, that folds back to 14 Khz due to us exceeding the system bandwidth and resulting aliasing. Therefore we go from a 21 Khz signal which most likely was not audible to a 14 Khz tone which may very well be. In analog world, we just have a 29 Khz tone which is an entirely different animal than 14 Khz. Yes, I made up this example on purpose :). But since you have no control over your music frequency content and that of jitter, it certainly can be one of the outcomes.

Summary

1.Yes, Flutter and Jitter are forms of FM modulation. For the reasons that it hurts differently if you got hit by a golf ball in the head than a basketball :D, just because the roots are the same, it doesn't mean perceptual effects are the same. Flutter amplitude is considerably higher and hence, creates a different situation.

2.Modulation index determines the bandwidth of an FM signal and flutter by definition has a wider distortion spectrum than jitter.

3.While frequency masking is alive and well at all times, its effect can sharply be reduced based on modulation index and with it, allowing distortions to be heard that would not otherwise be as audible.

4.Wow and Flutter standardized curves have a 4 Hz maximum importance with the amount dropping on both sides of that center frequency. Therefore Arny's original thesis is invalidated that as flutter frequencies increase, there is a downward slope in audibility of the distortion. He has acknowledged the same although it doesn't look like anyone picked up on that.

5.As I noted many pages back from the well written and simple presentation by FHG, the explanation of 4 Hz peak appears to be very related to temporal masking where the distance between fluctuations determines audibility.

6.There are cases such as jitter aliasing and data dependent jitter that simply have no analog in flutter. So as much as the levels of jitter can be lower than flutter, it can manifest itself in ways flutter cannot. So one has to be very careful in drawing parallels here. (There are other considerations here that I am not covering here just yet.)
----------------------------------
 
Nicely done. And that my friends is why this is such an important breakthrough if we do say so ourselves. The distortion reduction is important. That's exactly what was being "unmasked" in the Stevie Wonder example. I would imagine that many people don't hear a massive difference - we have to limit the action in the presence of noisy signal... and we did not expect much from this, we were lucky enough to find a small thread of bias in the 1/2 speed playback from one of Bruce's Studers. With the wideband head and preamp combo we would have done much better. But the comments above are exactly what we heard. There's also a musical and rhythmical solidity that is subtle but obvious once you're used to hearing it. The groove is tighter.
 

About us

  • What’s Best Forum is THE forum for high end audio, product reviews, advice and sharing experiences on the best of everything else. This is THE place where audiophiles and audio companies discuss vintage, contemporary and new audio products, music servers, music streamers, computer audio, digital-to-analog converters, turntables, phono stages, cartridges, reel-to-reel tape machines, speakers, headphones and tube and solid-state amplification. Founded in 2010 What’s Best Forum invites intelligent and courteous people of all interests and backgrounds to describe and discuss the best of everything. From beginners to life-long hobbyists to industry professionals, we enjoy learning about new things and meeting new people, and participating in spirited debates.

Quick Navigation

User Menu