Greenears
100+ Head-Fier
- Joined
- Dec 2, 2014
- Posts
- 248
- Likes
- 32
I'd be careful testing that one. What was the definition of instantaneous DR at a sample?
I'd be careful testing that one. What was the definition of instantaneous DR at a sample?
Thank you so much for the file - I downloaded it fine. Unfortunately on this network right now I (a) don't have a Unix prompt and (b) don't have gcc installed. I guess I could download it for Windows, or borrow a friend's system to try. Not a big barrier I'll try it at some point in the next couple of weeks.
I assume the 44 byte .wav format above is the standard one that comes out of SOX.
I get about 60dB or so for the the LOTR track (taking the 99th percentile of 2s RMS values and the max peak). I don't know how LKFS loudness works (I assume it's meant to work in concert with standards for delivery systems in theaters), but it seems like it means more like "you can't have a louder perceived sound than -31dB full scale" rather than a dynamic range measurement.
I think everyone needs to look at this post and be reminded of something key: All DR measurements require a window of time and a method of averaging (in this case time window is 2s and the method is 99 percentile etc) The instantaneous DR (1 sample time) can always be much higher as has been pointed out about 20 pages ago. I'm told that most modern tracks are normalized to peak around -6 or -3 dBFS you basically could have 90 dB instantaneous DR with redbook.
The real question I and others are asking is (a) how much of these instantaneous peaks are there (b) are there enough to become audible (c) can you tell the difference if you extend them to 120dB DR (20 bits equivalent). Another reference pointed out that a short peak of 120 dB is not noticeable, even though longer exposure (seconds?) will cause pain.
This updated version of the quantize program includes a Win32 executable, and supports 24-bit PCM samples (files created by sox and the dsputils programs should work). When processing 24-bit input, the output file is always in the same format, so it does waste some space. Another change is that adding 4 to the dither type switches from rounding towards the nearest integer to floor rounding.
And 1GHz is a much higher frequency than 20kHz, but so what? Your ears can't hear it and your ears also can't hear sounds at -144dB, which are represented by the 24th bit in digital audio. Your ears can't even hear the 18th bit. This is all wishful thinking, and proven as such in controlled tests.
And 1GHz is a much higher frequency than 20kHz, but so what? Your ears can't hear it and your ears also can't hear sounds at -144dB, which are represented by the 24th bit in digital audio. Your ears can't even hear the 18th bit. This is all wishful thinking, and proven as such in controlled tests.
Inaccuracies like these is what irritates me about this thread.
The original poster made the same mistake, claiming all there was to 24 vs. 16 bits was dynamical range.
Well from what it looks to me anyway. I'll try to explain, maybe I myself misunderstood something.
People here seem to confuse dB with dB-audio, the latter being an absolute measure (because it has an absolute reference, not depending on device), the former not.
dB just says how much louder it is compared to a reference level, the dynamic range of e.g. a soundcard being how much louder the loudest value is from the most quiet one.
But this could be scaled / mapped onto a differently spaced grid(e.g. when you fumble with the volume dial of your amp), not changing the dynamic range being used, as not only the loudest amplitude of a recording will be louder, but also the most quiet one (so the absolute audio level range changes, of course).
E.g.: if your DAC has a nominal dynamic range of 8bits or 48dB let's say outputting 1mV..255mV, and you externally amplify so that you get values from 25.5mV..6553.5mV, it's still 48dB, but your speakers put it into a different dbA range, and onto a different place within the human hearing range.
Nobody mentioned the one thing that would be deciding whether more bits can offer more perceived quality or not: The resolution of the ear itself. And it is indeed about resolution, not just range.
You could make a DAC which can produce only two different output levels, and space those such that they span 100 dB, so you have a "dynamic range of 100dB", but only two steps, which is obviously useless. It's about how many different amplitude values within the healthy dynamic range of the ear the latter can discern, similar to color formats' bit depths and the number of different shades of colors the eye can discern.
So more finely grained steps along the amplitude axis can matter, as long as it does not exceed the number of amplitude steps the ear can discern within the region of the overall hearing range of sound pressures that this granularity is mapped onto.
If, hypothetically, the ear could discern 1 bilion different audio levels within the 140 or so dB*Audio* level range of the ear, you could easily map those 144 dB range of a 24bit DAC onto, say, the center third of the ear's dbA range, and the ear could still discern all the steps - i.e. if there were less steps resolution there, the ear might notice the equivalent of visual banding in low number of color shades images!
Note that I am not making any specific claims about bit rates and their supposed advantages here.
Just generally speaking, the explanations given in this thread so far seem inaccurate.
From what I remember (can't find a reference) the ear actually has a resolution of about 17 bits, which would mean it could discern twice as many different audio levels than 16bit wave can. So 24 bits would be 128 times overkill - assuming that during playback it is all amplified perfectly into just the sound pressure range of hearing where it matters.
How this all really works out when people playback their music, I'm not sure ^^
.
[text]
.
[text]
.
As far as dynamic range, it is intimately related to the number of bits, as a lower number of bits simply cannot represent the same minimal and maximal waveforms that a higher number of bits can (techniques like noise-shaping notwithstanding). Take 16 vs. 8 bits and square waves. The maximal amplitude symmetric (around 0) square wave you can get at 16bits has its peaks at ±32767, while the minimal has peaks at ±1. That's a 90dB difference in amplitude. At 8bits, you get maximal peaks at {0,254} and minimal at {126,128}, which is a 42dB difference in amplitude. So we can get 48dB more difference between peak values by going to 16bit for this waveform.
Resolution, as you are using it, would come down to being able to detect the smallest change possible for a given signal. So for square waves that would mean, at 16bit, detecting the difference between ±32767 and ±32766. Maybe I'll make that example.
and I was afraid my explanation wouldn't be clear enough for him ^_^.
Inaccuracies like these is what irritates me about this thread.
The original poster made the same mistake, claiming all there was to 24 vs. 16 bits was dynamical range.
Well from what it looks to me anyway. I'll try to explain, maybe I myself misunderstood something.
People here seem to confuse dB with dB-audio, the latter being an absolute measure (because it has an absolute reference, not depending on device), the former not.
dB just says how much louder it is compared to a reference level, the dynamic range of e.g. a soundcard being how much louder the loudest value is from the most quiet one.
But this could be scaled / mapped onto a differently spaced grid(e.g. when you fumble with the volume dial of your amp), not changing the dynamic range being used, as not only the loudest amplitude of a recording will be louder, but also the most quiet one (so the absolute audio level range changes, of course).
E.g.: if your DAC has a nominal dynamic range of 8bits or 48dB let's say outputting 1mV..255mV, and you externally amplify so that you get values from 25.5mV..6553.5mV, it's still 48dB, but your speakers put it into a different dbA range, and onto a different place within the human hearing range.
Nobody mentioned the one thing that would be deciding whether more bits can offer more perceived quality or not: The resolution of the ear itself. And it is indeed about resolution, not just range.
You could make a DAC which can produce only two different output levels, and space those such that they span 100 dB, so you have a "dynamic range of 100dB", but only two steps, which is obviously useless. It's about how many different amplitude values within the healthy dynamic range of the ear the latter can discern, similar to color formats' bit depths and the number of different shades of colors the eye can discern.
So more finely grained steps along the amplitude axis can matter, as long as it does not exceed the number of amplitude steps the ear can discern within the region of the overall hearing range of sound pressures that this granularity is mapped onto.
If, hypothetically, the ear could discern 1 bilion different audio levels within the 140 or so dB*Audio* level range of the ear, you could easily map those 144 dB range of a 24bit DAC onto, say, the center third of the ear's dbA range, and the ear could still discern all the steps - i.e. if there were less steps resolution there, the ear might notice the equivalent of visual banding in low number of color shades images!
Note that I am not making any specific claims about bit rates and their supposed advantages here.
Just generally speaking, the explanations given in this thread so far seem inaccurate.
From what I remember (can't find a reference) the ear actually has a resolution of about 17 bits, which would mean it could discern twice as many different audio levels than 16bit wave can. So 24 bits would be 128 times overkill - assuming that during playback it is all amplified perfectly into just the sound pressure range of hearing where it matters.
How this all really works out when people playback their music, I'm not sure ^^
.
This updated version of the quantize program includes a Win32 executable, and supports 24-bit PCM samples (files created by sox and the dsputils programs should work). When processing 24-bit input, the output file is always in the same format, so it does waste some space. Another change is that adding 4 to the dither type switches from rounding towards the nearest integer to floor rounding.
You are theoretically correct about the distinction between resolution and DR. I used to try and explain this to photographers who thought that a 16-bit/channel image had higher DR than the same file converted to 8-bit. No, the DR was the same but the increments were less smooth.
But... all of this becomes moot as no proper double blind test has ever revealed a single test subject who could tell the difference between a live audio feed, and the same feed fed through a 16/44 AD-DA loop, which is evidence enough for me that although many problems exist in theory, in real life they are inaudible.
You are theoretically correct about the distinction between resolution and DR. I used to try and explain this to photographers who thought that a 16-bit/channel image had higher DR than the same file converted to 8-bit. No, the DR was the same but the increments were less smooth.
But... all of this becomes moot as no proper double blind test has ever revealed a single test subject who could tell the difference between a live audio feed, and the same feed fed through a 16/44 AD-DA loop, which is evidence enough for me that although many problems exist in theory, in real life they are inaudible.
All this back and forth on DR can be settled by this new utility posted. Simply keep chopping of a few bits at a time and till you pass the ABX test, then run the track through a DR analyzer.
I think people will find the standard equation of DR = N bits * 6 in dB only works at the limit where DR is at the max of N*6, but not at lower values. You'll be able to hear noise at levels lower than the equation predicts.
But .... I've been wrong before :--) Or have I?
Originally Posted by unshavenbastard /img/forum/go_quote.gif
You could make a DAC which can produce only two different output levels, and space those such that they span 100 dB, so you have a "dynamic range of 100dB", but only two steps, which is obviously useless. It's about how many different amplitude values within the healthy dynamic range of the ear the latter can discern, similar to color formats' bit depths and the number of different shades of colors the eye can discern.
Originally Posted by unshavenbastard /img/forum/go_quote.gif
From what I remember (can't find a reference) the ear actually has a resolution of about 17 bits.