[RWP] Latency, is shorter always better?
Chris Belle
cb1963 at sbcglobal.net
Tue Jul 22 21:42:08 EDT 2014
Not supprising with audacity since it does not support asio and you have
to manually set the offset to get tracks to play back in time.
On 7/22/2014 10:37 AM, Jim Snowbarger wrote:
> Tape decks with a moveable playback head? Very nice. I didn't know
> about that. But, that would have been great for adjusting the rate of
> regenerative feedback to match the tempo of a song.
> Another way to do that was to vary the speed, except that it had
> frequency response implications as well.
>
> An interesting experiment with DAWs, record a series of clicks into
> track 1. Arm track 2, and route the playback of track 1 into the input
> for track 2. Do it externally, rather than inside the computer, so you
> get the benefit of the entire chain of processing.
>
> Now, play tracks 1 and 2 back together. Listen for the time delay
> between the clicks on the two tracks. Ideally, they will be coincident.
> On reaper here, it isn't quite perfect, but it is pretty darned good.
> Last time I tried Audacity, they arrived in different time zones.
>
> ----- Original Message ----- From: "Chris Belle" <cb1963 at sbcglobal.net>
> To: "Reapers Without Peepers" <rwp at reaaccess.com>
> Sent: Tuesday, July 22, 2014 12:46 AM
> Subject: Re: [RWP] Latency, is shorter always better?
>
>
>> One more thought about this latency thing.
>>
>> And us old farts who used to play with tape decks will remember this.
>>
>> How about those 3 head decks where you could listen to play-back
>> while recording?
>>
>> YOU could hear the input while it was going down, and listen to your
>> playback head, and some of those were moveable.
>>
>> So you could change the latency between when something got recorded
>> and played back by moving the head closer or further away from the
>> recording head.
>>
>> That was on the commercial machines.
>>
>> I never had one of those, but i did have a very nice 3 head cassette
>> deck.
>>
>> This is, in fact, somewhat similar to what hapens in your daw,
>> even if your not recording but just listening to the signal comming
>> back from your daw once it goes through the internal processing, and
>> any plug-ins you might have,
>> and they add their own latency you can bet, and most modern daws
>> compensate for that under the hood.
>>
>> Sonar had automatic plug in delay compensation
>> way before many daws, including protools ever had it.
>>
>> Yes, you go back and listen to first episodes of the home recording
>> show on protools 9 and 10,
>> and you can hear them talking about lining up tracks manually after
>> the fact, to make the audio come out right, after going through all
>> the processing plugs.
>>
>> Boy howdy, now, isn't that a real pain in the posterior?
>>
>> INteligently keeping up with when in the time line a recording
>> starts, and how to play it precisely in the right way to account for
>> plug-ins latencies, and then play it properly again when you take
>> plugs of is not an easy task, but your daw does that all for you
>> under the hood, if it's worth a squat.
>>
>> People sometimes get in real trouble even with this automatic
>> stuff going on by not routing their monitoring right,
>> because there are certain ways of routing and recording which makes
>> it impossible for your daw to implement delay compensation properly.
>>
>> So this is why I tend to like to not do plugs until after I've laid
>> my audio.
>>
>> Not always possible, you can't lay that heavy rock guitar track
>> easily only hearing plink, plunk, twang,
>> but you can believe your daw is doing the latency shuffle dance when
>> you have many tracks playing and you are laying guitar amp simms
>> which has latency going both ways, because remember, you are going
>> audio in, and audio back out,
>> and this is why with mastering plugs which cause a lot of latency,
>> especially multi-band compressors with look ahead and back in my
>> early early days of learning this stuff it used to drive me nuts, why
>> are my midi tracks being delayed so much when I press a note but they
>> play just fine on playback?
>>
>> Well, it's that delay compensation working for you.
>>
>> Imagine having to figure out how much delay you had and fixing all
>> that manually?
>>
>> You can get interesting things happening when using reverb in
>> projects by turning off delay compensation,
>> you get a built in pre-delay, which is a setting on high quality
>> reverb units, the reverb doesn't start right away, and
>> this helps make room in the mix when you don't wan the verb in the way,
>> and it kicks in after the initial atack of your audio.
>>
>> Or do we remember real world latency,
>> and the days when distructive editing was the only kind you did, if
>> you wanted to process an equalizer, or chorus fx, you hit the button,
>> and then go have a sandwich and waited for your 486 to process that
>> track, and you'd come back 10 minutes later and maybe have a wet and
>> dry track.
>> and you could do interesting things with that by time delaying the
>> wet track 'grin'.
>>
>> When I do drum replacement by generating midi tracks from transient
>> points of an audio drum track and then feeding it to a audio bus with
>> samples, I have to time align the new track to match the old one,
>> at least in the old days we had to do more of that before delay
>> compensation was automatic.
>> in most daws.
>>
>> Still, most daws will only do this in a certain range, see above,
>> where I mention mastering plugs,
>> linear phase equalizers are also notorious for introducing way too
>> much delay to use them in real time.
>>
>> So are transient processors, shapers.
>>
>> Maybe when we get processors running at 30 gigahertz
>> we'll be able to do that stuff in real time, and did I hear silly
>> people want to make a daw out of an ipad?
>>
>> Right now in 2014, an ipad will just barely run a guitar simm with
>> low enough latency
>> \to be playable.
>>
>> Well, what do you expect from a little baby toy computer?
>>
>>
>> On 7/21/2014 9:33 PM, Jim Snowbarger wrote:
>>> Now and then, I feel like a slight departure from topic.. And, this
>>> is one of them. So, stand bye with your delete key ready as I carry
>>> on.
>>>
>>> This probably belongs over on MidiMag. But, I don't feel like
>>> joining just so I can post this once in a blue mooner.
>>>
>>> One of the great things that digital audio processing has brought to
>>> us is so-called latency. You might just call it delay. but, in the
>>> 21st century, we like to use clever names. It makes us feel
>>> smarter. So, let's co-opt the term latency, which had a totally
>>> different meaning before the techno-gods got hold of it. And, let's
>>> now define latency as the act of being late. But, however you slice
>>> it, it comes down to delay.
>>>
>>> Digital devices impose delay mostly because data consumers, like
>>> sound cards, or recording devices, have learned to be defensive,
>>> knowing full good and well that data providers, such as input sound
>>> cards, or other streaming devices, can not be counted on to keep up
>>> a steady stream of data. Internet congestion, or scheduling
>>> congestion inside your own machine, can temporarily block the normal
>>> flow of things. Sound playback requires a rock-solid comsumption
>>> rate of the data. The sampels need to keep flowing. You might not
>>> get that next buffer load of data in time. so, it pays to keep a
>>> backlog. The more backlog, the safer you are. But, if the backlog is
>>> too great, you get, latency, that annoying delay.
>>>
>>> I recently picked up one of those fine Computers Chris is always
>>> talking about from StudioCat.com. That is one very fine box. And,
>>> now that I also own Chris's Delta 1010, I was enjoying fine-tuning
>>> my latency down to acceptable levels, not carefully measured, but
>>> clearly less than 10 milliseconds.
>>>
>>> Most of the recording work I do involves a microphone and
>>> headphones. I am quite typically listening to my own voice as I
>>> speak. If you have listened to the Snowman Radio Broadcasts, you
>>> know the kind of multi-track microphone work I'm guilty of.
>>> When living on machines where such short delays were not possible,
>>> my habit was to listen to my own foice direct out of the mixer, and
>>> not going through Reaper. So, I kept the reaper monitor off. What
>>> was annoying about that is that, if I panned my various character
>>> voices in the stereo mix, then, my direct microphone sound would
>>> not be panned the same as the character voice track I was recording
>>> into. So, when it played back, it came from elsewhere, and was more
>>> than a little bit confusing.
>>>
>>> But, with delay this short, I find that I switch off the direct
>>> sound, and now can monitor the signal coming back from reaper with
>>> the monitor turned on. So, I'm now listening to a delayed version
>>> of my voice, and it is panned to the same place where that character
>>> voice sits, which helps me keep track of who I am supposed to be
>>> right now. And, I can more easily tell now whetehr a track is
>>> armed, and even if one is armed that should not be. It's nice to be
>>> able to work like that, just listening to reaper's output.
>>>
>>> But, here is the cool thing. The exact amount of latency you
>>> provide affects the quality of what you hear in your headphones.
>>>
>>> No matter how good your phones, the sound that you hear when you are
>>> listening to yourself speaking live into a microphone, is actually
>>> the composite of at least two signal paths, and maybe more. Yes,
>>> there is the direct signal coming through Reaper. Then, there is
>>> bone conductivity, the sound of your own voice coming through the
>>> structure of your head, which will very somewhat with density. If
>>> you don't get any of that, you might wonder about that density stuff.
>>> And maybe even, there is leakage around the ear muffs. In all, it
>>> is a complex sond that actually reaches your ears. And, the phase
>>> relationship between all of the various contributors will affect the
>>> frequency response of the final signal that you hear.
>>>
>>> In the old days, we knew about the affect that phase would have on
>>> such things. Having your head phones out of phase with your
>>> microphone left you feeling empty headed, due to the phase
>>> cancellation that took place.
>>> But, since delay was in the nanoseconds, we didn't get to know so
>>> much about the effect that delay would have, despite our compulsive
>>> preoccupation with tape delay.
>>>
>>> Phase is mostly a frequency independent phenomenon. Yes, we know
>>> that some systems, especially mechanical transducers, or even cheap
>>> equalizers, which will have a reactive component to their impedance,
>>> introduce a variable amound of phase shift, depending on frequency.
>>> But, usually those effects are at the far ends of their usable range.
>>> In general, especially in mixer land, where things are nice and
>>> linear, and where impedances are strictly non-reactive, if you put
>>> something 180 degrees out of phase, you will get perfect
>>> cancellation, all across the frequency band.
>>>
>>> Enter the digital age, and the new innovation, latency.
>>> The relationship between signal phase, and a delay is frequency. For
>>> example, a delay of 4 milliseconds is one full cycle of a 250 Hertz
>>> tone. But, it is only half a cycle of a 125 hertz tone. It is all
>>> still a 4 millisecond delay. But, the phase impact depends on the
>>> frequency. Combining the pre and post delays of these two tones with
>>> that 4ms delay will have completely different effects. The 125 hertz
>>> tone would be nulled out. The 250 hertz tone would actually see a 6
>>> db increase.
>>>
>>>
>>> The result is that, if you put a delay in front of your headphone
>>> mix, you will cause what is referred to as a comb filter effect on
>>> the perceived headphone signal. It is a filter that has a frequency
>>> response curve that looks like a rola coaster, with hills and
>>> valleys. If you were listening to an audio tone sweep, one that you
>>> would actually need to sing, in this case, in order to get that bone
>>> conductivity thing happening as well, As you move steadily up in
>>> frequency, the sound would be much stronger at some frequencies, and
>>> much weaker at others. As the tone rises, you would hear rising and
>>> falling of the net response. And, changing the amounbt of delay
>>> slides that comb up and down the audio spectrum.
>>> Depending on several things, the frequency range of your voice,
>>> response of your headphones, your ears, the density of your grey
>>> matter, your preferences, and on and on, you might have preferences
>>> about the optimal position of that comb. What frequencies do you
>>> like to accentuate? And, which to attenuate.
>>>
>>> The cool thing is that, by fine-tuning your headphone latency, you
>>> can position that comb how you like, and can optimize your headphone
>>> experience. The latency needs to be short enough to not give you a
>>> delay echo effect. But, beyond that, the shortest possible latency
>>> may not give you the headphone experience you like. Instead, relax
>>> it a little, and see what enriching tones come your way.
>>> Silly you. And you always thought shorter was better. And now you
>>> know.
>>> TROTS.
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> RWP mailing list
>>> RWP at reaaccess.com
>>> http://reaaccess.com/mailman/listinfo/rwp_reaaccess.com
>>>
>>
>>
>> _______________________________________________
>> RWP mailing list
>> RWP at reaaccess.com
>> http://reaaccess.com/mailman/listinfo/rwp_reaaccess.com
>>
>
>
>
> _______________________________________________
> RWP mailing list
> RWP at reaaccess.com
> http://reaaccess.com/mailman/listinfo/rwp_reaaccess.com
>
More information about the Rwp
mailing list