blairl wrote on Tue, 12 October 2004 05:38 |
One of the major concerns on the thread in George's forum is that of monitoring during the tracking and overdub process. Let's take a vocalist for example that is overdubbing to prerecorded tracks. Since a vocalist can hear their own voice in their head, some complain that the delay of microphone to A/D - D/A conversion and back through the headphones is unnatural and irritating. Moving the microphone closer still won't compensate fully since the voice is already present in the head. Some vocalists reportedly have a hard time adapting to hearing their voice delayed. At 44.1 and 48K the predominant DAW interface has a latency of approximately 2ms A/D - D/A.
Have you done any studies on how latency affects monitoring in the recording process? Do you know of any limits before latency starts to affect the performance?
You mentioned alternate converter chips for lower latency. Would using these chips degrade the sound quality in favor of lower latency?
|
“Since a vocalist can hear their own voice in their head… some complain that the delay of microphone to A/D - D/A conversion and back through the headphones is unnatural and irritating.”
I understood that, and that is what we call latency.
“Moving the microphone closer still won't compensate fully since the voice is already present in the head.”
What do you mean when you say won't compensate fully? There is going to be a delay due to many factors. Some factors contribute little (add little delay) and others add a lot of delay. Let’s look at the proposed desired latency of 500usec. Let us assume that the mic itself contributes zero, the speaker (or headphone) contributes zero, the electronics in series contribute zero. Than there is NO WAY you can get 500usec latency when the distance between mic and vocal chord is much greater than 6 inches!
So I assume the comment regarding 500uSec was for the electrical portion only (AD and DA). But think about it - an electrical path of 500uSec and 6 inches acoustic distance is still very restrictive.
I am not saying that one can not hear tracks moved relative to each other by 500uSec or by 1usec. I am not talking about what one can hear or not hear. What I am suggesting that we all may have to learn to live within some physical limitations, and “mentally adjust” for more delay. The recording engineer can later slide tracks to make it the way they want it to be.
“Some vocalists reportedly have a hard time adapting to hearing their voice delayed. At 44.1 and 48K the predominant DAW interface has a latency of approximately 2ms A/D - D/A.”
So with 2msec AD and DA, you add the acoustic delay (1msec for 1 foot, 2msec for 2 feet…) and there you are at some latency. If it is too much, you cut the delay down but do it wisely, not by trying to convert a car into a jet. The sigma delta technology is a “car” (when talking about latency). Another architecture is a “jet” for latency. Proposing 384fs to reduce latency is an attempt to convert all the cars to jets. It is the difficult way to do things and everyone has to buy a expansive jet. No more cars. Jets are difficult to park, they take a lot of fuel…
”Have you done any studies on how latency affects monitoring in the recording process? Do you know of any limits before latency starts to affect the performance?”
My first message on the forum (see comments) states that we will not deal with ear brain issues. I rather stay in my domain.
“You mentioned alternate converter chips for lower latency. Would using these chips degrade the sound quality in favor of lower latency?”
Clearly, latency is not the first in the list of what is important in audio. If “latency
”ruled, the IC makers would not have invested so much in the popular sigma delta architecture, which is in fact, takes time to convert. Most of the delay in an AD is due to a decimation circuit, and in a DA it is due to the up sampling circuit. These digital circuits FIR type filters (finite impulse response) are used because they provide a property called “linear phase”.
Recently, an IC maker (Cirrus, formerly Crystal Semiconductors) has introduced a number of IC’s (both AD and DA) that utilizes a different digital filter structure called IIR (infinite impulse response). The IIR filter is very fast (small delay), but it does not yield linear phase. It may be a great solution for the specialty type of application we are talking about. Other than the linear phase tradeoff, those IC’s are very good performers.
Again, it is up to the ear people to tell us the EE’s how much deviation from linear phase is acceptable. Knowing the parameters such as latency, phase linearity and more, we can make progress. At this point, latency oriented gear is a “specialty market”.
Roughly speaking, with 500usec “budget”, analog seems like the wisest solution. Every 100usec’s is an inch! With 3-4msec “budget” you can do fine with what is on the market. If you are going to insist on digital in the range of say 500uSec to 2msec (electric plus acoustic delay) you open up a “specialty market”. Such gear will be a bit more costly (less volume). Quality? Again, roughly speaking, you can have better than CD quality today, at reasonable cost. Shooting from the hip (but with a lot of converter design experience) I would settle on 88.2-96KHz as the optimum rate, for what I think is a good compromise.
Again, with such a tight requirenment for latency, I would consider analog the best solution.
BR
Dan Lavry