71 dB
Headphoneus Supremus
Can you please dial down the drama.
Already feeling better so yes, I can. I'm done with µ-law.
Can you please dial down the drama.
[1] In perceptual coding, audio data is selectively removed (compressed) based on how unlikely it is that a listener will notice the removal. ...That [definition] is true for modern codecs, but it is not a general rule/definition.
[2] In your exchange with ILoveMusic, there are several ideas that are unclear, misleading, garbled or incorrect. Not all of that from you(!) in the above quote(!!), but rather in the entire exchange.
If you’re so inclined, perhaps you can comment on the following facts:
1. That was exactly my point. All that is actually measured/stored in the digital domain is amplitude vs. time. Frequency is derived from that "amplitude vs time" information, so effectively there is only one domain "amplitude vs time".1. ... The analog signal from a microphone, the signal on an analog interconnect, the output of a DAC or the signal on the speaker wire from an analog amplifier are all voltage amplitude vs. time (time domain). Frequency information is not available unless one transforms the signal, using a spectrum analyzer for analog data or a Fourier transform for digital data.
2. 16 bit in not perfect. [2a] If I have an original signal, convert it to 16 bit resolution, and use that to create a reproduced signal, the original and reproduced will not be identical. That is, subtracting the two does not give all zeroes.
3. 1 bit delta-sigma coding (used in DSD and SACD) is not the same as 1 bit LPCM coding. Usually, talking about 16 bits implies linear pulse code modulation.
4. Shannon-Nyquist tells us we need to sample at greater than twice the highest frequency of interest, not greater than or equal to twice. Twice the highest frequency is inadequate.
[1] That is why, since the early 1990s I guess, production has moved from 16/44.1/-96 to 24, 32-float/44.1, 88.2, 96, 192/-144 to -infinity dynamic range, instead of remaining at the same specs as Redbook, as it did through most of the eighties.
[2] Every time processing is done, in the digital domain, quantization occurs, and thus dither must be applied.
[1] But bit-depth(8, 16, 24, etc) does determine the amount of quantization error/noise. So even if conditions(within the bandwidth of the sampling rate) are met, there will always be 'noise'. 48dB down from full-scale for 8bit, 96dB down from full-scale for 16bit, and so on.
[2] 96dB dynamic range? Under any conditions? You're simply wrong. Upon what do you base your statement?
1. Not really. In the '80's the vast majority of production (mixing) was still analogue, even though the recording was often digital. In the 1990's recording progressed from 16bit to 24bit, not because 24bit had more accuracy/sound quality than 16bit but simply because it provided far more headroom, making multi-channel recordings much easier.
2. No, that's not correct. Yes, every time processing is applied in the digital domain quantisation error occurs but dither does NOT need to be applied. In the '90's, commercial digital mix environments were usually 48bit fixed or 32 float. Many hundreds of processing steps would be required for the cumulative quantisation error to get even near audibility and therefore dither is not typically required. Today (and for quite a few years), commercial mix environments are virtually all 64bit float and the quantisation error is so tiny, you'd need to sum together thousands of processing steps to even get within the theoretical limits of 24bit!
1. Mmmm, that's not exactly correct. If it were correct, 1bit (SACD) would be un-listenable with only 6dB of dynamic range. Of course, SACD is very listenable, the noise is effectively a great deal further down than -6dB, due to noise-shaped dither. Therefore:
2. Even in most top class commercial recording studios a 96dB dynamic range is typically more than uncomfortable. With a studio noise floor of say 20dB, a 96dB range above that would put the peaks at 116dB which is not far off the threshold of pain, let alone "comfortable". However "96dB dynamic range" is rather arbitrary, referring to the point above, the dynamic range of CD/16bit is not 96dB, it's effectively about 120dB, due to noise-shaped dither. Can you think of any conditions where a dynamic range of 120dB would be comfortable? In practice, virtually all commercial audio is kept within a 60dB dynamic range, as more than that would be uncomfortable for the majority of consumers.
G
(...) Anyone...even you!....could write an article. (...)
most of us understand that there isn't much of anything perfect going on anywhere from recording down to playback output.
(...)
does this make sense, or did I just enter the twilight zone in my own head?
Well, when I found it here, I thought it caught the gist of the idea, but you have indeed found a loophole. But your loophole does not only use the "perceptual" (noise-shaped dither) part of the definition, your argument includes the idea that RBCDs use "lossy compression". It it were 2 in the morning in the dorms, the pizza just arrived, and we just reloaded the bong or grabbed a six-pack, I'd be up for arguing for hours whether down-sampling and truncation to create CDs constitutes "lossy compression". But I moved out of the dorms 40 years ago and haven't "smoked" in nearly as long, so I'm not interested in this idea.That definition doesn't work though. For example, reducing a 24bit signal down to 16bit with the application of noise-shaped dither would also qualify as "perceptual coding" and therefore most CDs produced in the last 20 years or so are "perceptual coding". In modern usage, the term "perceptual coding" is used specifically as I stated and does not include redbook CDs.
Cool, that's a neat trick. Indeed, I didn't read that part. If you put that in your sig, you never have to worry about being wrong again. You just say "Dude, read my sig". Sweet.I made it clear a number of times that I was not attempting to be absolutely accurate, instead, I was looking for a view or "visualization" which would help him understand some of the fundamental basics. For example, in that same post I stated "BTW, all the above is not exactly correct or incorrect either! It's just another way of looking at the issue, a way which avoids some incorrect conclusions/assumptions."
1a. I don't even get how you can make this mistake. Freq. and time domains are 2 domains. There are complete textbooks on freq. domain analysis, that don't work in the time domain, and vice versa. I mentioned convolution vs. point-by-point multiply... they are different.1. That was exactly my point. All that is actually measured/stored in the digital domain is amplitude vs. time.
1a.Frequency is derived from that "amplitude vs time" information, so effectively there is only one domain "amplitude vs time".
among so many dozens more over the years. We don't measure/convert frequency. We measure/convert amplitude as a function of the independent variable, time.Since the dawn of audio recording right up to the present day we measure/convert and store just two properties, frequency and amplitude.
2. No, BigShot is right. Shannon-Nyquist deals with sample rate and bandwidth, not bit depth. (@bigshot-I never claimed Nyquist deals with bit depth). You have mistaken the complete lack of bit depth in Shannon-Nyquist to mean bit depth is irrelevant. That is not what it means. Sampling theory deals with periodically taking samples... but garbage-in, garbage-out... the quality (resolution) of the sample determines the quality of the reconstruction. Perfect reconstruction requires perfect samples. I think BigShot may want to add: "But good enough is good enough". Perhaps a dictionary is needed to agree on the meaning of "perfect"! ...?2. 16bit is perfect, any bit depth it's perfect and to suggest otherwise requires disproving the Nyquist/Shannon Theorem. However ..
2a. Two points: Firstly, if you convert your original signal to 16bits and then reproduce that signal you are going to need an Analogue to Digital Converter and a Digital to Analogue Converter, both of which, by definition, are not just digital devices but analogue devices as well. If you compare an input signal to an output signal you are therefore not only evaluating 16bit digital but 16bit plus two analogue stages, neither of which can be absolutely perfect.
I understand how some find weird that @gregorio would insist on some perfect aspects of digital audio so often while it's obvious that he himself recognizes the limits of a practical application as he just did. but here is why I believe he does that and is right to do so (maybe I'm wrong, he'll tell me^_^) ...
[1] Man you just love saying "No", "Not correct". [1a] Maybe I just got the timelines a little wrong??
[2] Well, the information I posted, that you claim is "not correct", is being retransmitted all over the web, on my Facebook pages and blogs.
[2a] And guess what Gregorio: I'm getting people to change both how they buy music and what they listen to it on!
[3] They, like I, are tired of having our chains jerked.
In short, Shannon's statement is logically incontestable, effectively it's an unquestionable absolute truth/fact. It's vital to fully appreciate this, Shannon's statement does not require any additional conditions in order to be true, it is true, period!! For example, note that it does not mention bit depth, that's because it does not need to, Shannon's statement is correct as it stands, period (IE. Regardless of bit depth). In effect, all "bit depth" does is define the amount of noise which accompanies our "complete determination" but the "complete determination" is always there and to suggest otherwise effectively means disproving what Shannon has already proven.
2. No, BigShot is right. Shannon-Nyquist deals with sample rate and bandwidth, not bit depth. (@bigshot-I never claimed Nyquist deals with bit depth). You have mistaken the complete lack of bit depth in Shannon-Nyquist to mean bit depth is irrelevant. That is not what it means. Sampling theory deals with periodically taking samples... but garbage-in, garbage-out... the quality (resolution) of the sample determines the quality of the reconstruction. Perfect reconstruction requires perfect samples. I think BigShot may want to add: "But good enough is good enough". Perhaps a dictionary is needed to agree on the meaning of "perfect"! ...?.
At 4-bit there is about 24 dB of dynamic range, so of course all the stuff below that is likely to be lost in the noise. So, what are these important features? However, it's perfect signal + really loud noise.-I will make a signal with a bandwidth less than 20kHz and duration of 1 second.
-You tell me exactly how you want me to prepare it, consistent with a simple test of your contention that bit depth doesn't alter the ability for "complete determination". Anything you want, ending with me sending you a file that is encoded at 4-bits and sampled at 44.1kHz. I will document each step with an intermediate, modified data file.
-You recreate my signal, showing a "complete determination" of my signal. If some of the important features of my signal areburied in noisemissing, that is not a success.
-We post the results here.
-If you do succeed... what should we do? Have a bet? You succeed, I send you money; you don't succeed, you send me money? Or we send to a charity of the winner's choice? Or? How much? 10, 100, 1000? Euros, dollars, pounds?
I'm confident. Are you?
[1] But your loophole does not only use the "perceptual" (noise-shaped dither) part of the definition, your argument includes the idea that RBCDs use "lossy compression".
[2] Cool, that's a neat trick. [2a] Indeed, I didn't read that part. [2b] If you put that in your sig, you never have to worry about being wrong again.
1a. I don't even get how you can make this mistake.
1. Yay! I'm so glad you have this brand new perspective that corrects so very many of your previous posts, e.g.
among so many dozens more over the years. We don't measure/convert frequency. We measure/convert amplitude as a function of the independent variable, time. ...
[2] As for helping "him understand some of the fundamental basics", how can you explain the need for anti-aliasing filters, the idea of frequency folding, the need for anti-imaging (reconstruction) filters, even really fundamental stuff like bandwidth and frequency response without the frequency domain?
2. No, BigShot is right. Shannon-Nyquist deals with sample rate and bandwidth, not bit depth.
2a. I was actually thinking of a double- or extended-precision float signal. Convert it to 16-bit integer (scaling properly, of course), and then convert back to the same floats, dither, filter, scale as needed, and the signal will be different.
[3] But you say any bit depth is perfect... 1 or 2 bit LPCM will really suck...