[RWP] Latency, is shorter always better?
Jim Snowbarger
Snowman at SnowmanRadio.com
Mon Jul 21 22:33:01 EDT 2014
Now and then, I feel like a slight departure from topic.. And, this is one
of them. So, stand bye with your delete key ready as I carry on.
This probably belongs over on MidiMag. But, I don't feel like joining just
so I can post this once in a blue mooner.
One of the great things that digital audio processing has brought to us is
so-called latency. You might just call it delay. but, in the 21st century,
we like to use clever names. It makes us feel smarter. So, let's co-opt
the term latency, which had a totally different meaning before the
techno-gods got hold of it. And, let's now define latency as the act of
being late. But, however you slice it, it comes down to delay.
Digital devices impose delay mostly because data consumers, like sound
cards, or recording devices, have learned to be defensive, knowing full good
and well that data providers, such as input sound cards, or other streaming
devices, can not be counted on to keep up a steady stream of data. Internet
congestion, or scheduling congestion inside your own machine, can
temporarily block the normal flow of things. Sound playback requires a
rock-solid comsumption rate of the data. The sampels need to keep flowing.
You might not get that next buffer load of data in time. so, it pays to
keep a backlog. The more backlog, the safer you are. But, if the backlog
is too great, you get, latency, that annoying delay.
I recently picked up one of those fine Computers Chris is always talking
about from StudioCat.com. That is one very fine box. And, now that I also
own Chris's Delta 1010, I was enjoying fine-tuning my latency down to
acceptable levels, not carefully measured, but clearly less than 10
milliseconds.
Most of the recording work I do involves a microphone and headphones. I am
quite typically listening to my own voice as I speak. If you have listened
to the Snowman Radio Broadcasts, you know the kind of multi-track microphone
work I'm guilty of.
When living on machines where such short delays were not possible, my habit
was to listen to my own foice direct out of the mixer, and not going through
Reaper. So, I kept the reaper monitor off. What was annoying about that is
that, if I panned my various character voices in the stereo mix, then, my
direct microphone sound would not be panned the same as the character voice
track I was recording into. So, when it played back, it came from
elsewhere, and was more than a little bit confusing.
But, with delay this short, I find that I switch off the direct sound, and
now can monitor the signal coming back from reaper with the monitor turned
on. So, I'm now listening to a delayed version of my voice, and it is
panned to the same place where that character voice sits, which helps me
keep track of who I am supposed to be right now. And, I can more easily
tell now whetehr a track is armed, and even if one is armed that should not
be. It's nice to be able to work like that, just listening to reaper's
output.
But, here is the cool thing. The exact amount of latency you provide
affects the quality of what you hear in your headphones.
No matter how good your phones, the sound that you hear when you are
listening to yourself speaking live into a microphone, is actually the
composite of at least two signal paths, and maybe more. Yes, there is the
direct signal coming through Reaper. Then, there is bone conductivity, the
sound of your own voice coming through the structure of your head, which
will very somewhat with density. If you don't get any of that, you might
wonder about that density stuff.
And maybe even, there is leakage around the ear muffs. In all, it is a
complex sond that actually reaches your ears. And, the phase relationship
between all of the various contributors will affect the frequency response
of the final signal that you hear.
In the old days, we knew about the affect that phase would have on such
things. Having your head phones out of phase with your microphone left you
feeling empty headed, due to the phase cancellation that took place.
But, since delay was in the nanoseconds, we didn't get to know so much about
the effect that delay would have, despite our compulsive preoccupation with
tape delay.
Phase is mostly a frequency independent phenomenon. Yes, we know that some
systems, especially mechanical transducers, or even cheap equalizers, which
will have a reactive component to their impedance, introduce a variable
amound of phase shift, depending on frequency. But, usually those effects
are at the far ends of their usable range.
In general, especially in mixer land, where things are nice and linear, and
where impedances are strictly non-reactive, if you put something 180 degrees
out of phase, you will get perfect cancellation, all across the frequency
band.
Enter the digital age, and the new innovation, latency.
The relationship between signal phase, and a delay is frequency. For
example, a delay of 4 milliseconds is one full cycle of a 250 Hertz tone.
But, it is only half a cycle of a 125 hertz tone. It is all still a 4
millisecond delay. But, the phase impact depends on the frequency.
Combining the pre and post delays of these two tones with that 4ms delay
will have completely different effects. The 125 hertz tone would be nulled
out. The 250 hertz tone would actually see a 6 db increase.
The result is that, if you put a delay in front of your headphone mix, you
will cause what is referred to as a comb filter effect on the perceived
headphone signal. It is a filter that has a frequency response curve that
looks like a rola coaster, with hills and valleys. If you were listening to
an audio tone sweep, one that you would actually need to sing, in this case,
in order to get that bone conductivity thing happening as well, As you move
steadily up in frequency, the sound would be much stronger at some
frequencies, and much weaker at others. As the tone rises, you would hear
rising and falling of the net response. And, changing the amounbt of delay
slides that comb up and down the audio spectrum.
Depending on several things, the frequency range of your voice, response of
your headphones, your ears, the density of your grey matter, your
preferences, and on and on, you might have preferences about the optimal
position of that comb. What frequencies do you like to accentuate? And,
which to attenuate.
The cool thing is that, by fine-tuning your headphone latency, you can
position that comb how you like, and can optimize your headphone experience.
The latency needs to be short enough to not give you a delay echo effect.
But, beyond that, the shortest possible latency may not give you the
headphone experience you like. Instead, relax it a little, and see what
enriching tones come your way.
Silly you. And you always thought shorter was better. And now you know.
TROTS.
More information about the Rwp
mailing list