Well, it’s been a busy weekend and this week has completely gotten away from me so far, but I’ll attempt to cover some of the points raised:
Firstly, there seems to be a misunderstanding about what we are trying to achieve – we are not trying to “fix” R-2R, or “fix” Delta/Sigma – we are trying to make the best DAC for audio that we can – and due to the high bar we’ve been setting since the early 90’s, there is an element of marginal gains at play – but why shouldn’t we continue striving to enhance performance?
To improve the state of the art you need subject matter expertise. Only by understanding the benefits and limitations of various approaches to A-D and D-A conversion can we design a DAC that we feel is best in class. Of course what we deem as a critical performance requirement may not be the same as other manufacturers or hobbyist engineers, however we are one of a select few manufacturers who have designed world-class ADCs, DACs and sample rate converters for both studio and consumer use.
So, we are trying to point out the issues that exist in common architectures, which to sum up are:
You can have lots of weighted current sources at a low sample rate – the challenge here is matching these current sources, and keeping them matched over temperature variations and time. For audio, the side-effects of this is that any errors in this matching cause unnatural distortion (due to correlation), which the human auditory system is very sensitive to. On the plus side, because we don’t have to run that fast, jitter is less of an issue.
You can have conceptually a single current source, and run it at a much higher sample rate. This fixes the matching issue (because it’s self-referencing – on or off, and any drift will manifest as DC rather than distortion). Unfortunately to deal with the quantisation noise generated, you have to heavily noise shape this and move it up in band. This can cause issues because if you keep the clock at sensible rates, the quantisation noise is very close to the audio band, and if you move it too high in frequency jitter becomes a real issue (due to switching noise), at which point you may have to perform some quite horrible (and sometimes impossible) maths to match rates.
What the Ring DAC does is effectively a hybrid – the clock can run at sensible rates (so 3-6MHz) and the noise shaping can be gentle, but because we have multiple codes to represent, we need a way to match them exactly - whilst bearing in mind that components age, temperature can become a factor and so on. This is the job of the mapper, and it has numerous attractions, including distributing DAC errors away from where we are interested (audio frequencies) to where we are insensitive (very high frequencies), without altering the data presented to it, whilst at the same time ensuring all the sensitive components age in the same way.
We can definitively state the that the Ring technology is not multiple DSD streams, and is not random – if you read the writeup, we even say “may appear random”.
It is quite correct to say that you cannot decorrelate noise that is already part of the signal. However, it is worth thinking about this as a philosophical point. One view is that noise shaping and filtering are evil because they somehow ‘guess’ and don’t reproduce the ‘original’ signal, and the ‘fewest steps must be the best’. Now, this may sound strange coming from us, but one of our beliefs can be attributed to Einstein – “Everything should be made as simple as possible, but no simpler”. So what is the original signal we are trying to reproduce?
We’ll cover this more in the filtering article, but here is a genuine question – on the scope shots supplied, we can see some effects that are clearly the lack of sensible filtering – if this is meant to show averaging, then the steps are way, way too big (consider all the discussions on matching have been < 1%). Does anybody think that the stair steps seen in the photographs are beneficial in any way? If so, how? Where do they come from?