TheSonicTruth
1000+ Head-Fier
- Joined
- Dec 19, 2014
- Posts
- 1,018
- Likes
- 126
No i dont sing for crap.
Hey I resemble that statement! I like R&B, among many genres.
No i dont sing for crap.
For simplicity PCM audio is linear. Each bit increases dynamic range by 6 dB. There is no need to increace dynamic range of 16 bit audio (16*6 dB = 96 dB). With dither it is more than enough (perceptual dynamic range up to 120 dB!). 24 bit (24*6 dB = 144 dB) is good in music production, but not needed in consumer audio. Even at studio all of that 24 bit dynamic range can't be used, because all electrical devices have higher background noise level.This article makes it sound to me that when using 24 bit audio instead of 16 bit, the range of loudness that the audio file can describe becomes greater. Why couldn't it instead increase the number of digital increments describing the difference between say -4db and -3db?
additionally, why couldn't the dynamic range of 16 bit audio be expanded by reducing the number of digital increments per 1db change?
Is the difference in db between the integer value that the 16 or 24 bits convert to a fixed value?
To me it seems wasteful to allow a 16 bit audio file to have a noise floor ~60db below the quietest noise in the music (assuming a song with 36db of dynamic range). Why can't the lowest noise or output voltage (-36db) be set to 0000 0000 0000 0001 and the highest noise (0db) to 1111 1111 1111 1111?
I apologize in advance for posting to such an old forum post. I yearn for a greater understanding of this subject.
Thanks in advance!
Good explanation. PCM is linear.For simplicity PCM audio is linear. Each bit increases dynamic range by 6 dB. There is no need to increace dynamic range of 16 bit audio (16*6 dB = 96 dB). With dither it is more than enough (perceptual dynamic range up to 120 dB!). 24 bit (24*6 dB = 144 dB) is good in music production, but not needed in consumer audio. Even at studio all of that 24 bit dynamic range can't be used, because all electrical devices have higher background noise level.
Remember, half of the 16 bits is used for negative values of the signal, so the maximum amplitude is 2^15 = 32768 = 0 dBFS. Then 23198 = -3 dBFS and 20675 = -4 dBFS. So, there are 2523 "increments" between -4 dBFS and -3 dBFS. This number of increments isn't that interesting. The level of quantization noise/dither is what matters.
There is µ-law/A-law 8 bit audio (companding algorithm) used in digital telecommunication that does what you propose. It results in quantization noise that fluctuates with the signal. The louder signal, the louder noise and vice versa. This was an early form of perceptual audio encoding.
[1] This article makes it sound to me that when using 24 bit audio instead of 16 bit, the range of loudness that the audio file can describe becomes greater.
[2] Why couldn't it instead increase the number of digital increments describing the difference between say -4db and -3db?
[3] additionally, why couldn't the dynamic range of 16 bit audio be expanded by reducing the number of digital increments per 1db change?
[3a] Is the difference in db between the integer value that the 16 or 24 bits convert to a fixed value?
[3b] To me it seems wasteful to allow a 16 bit audio file to have a noise floor ~60db below the quietest noise in the music (assuming a song with 36db of dynamic range).
[3c] Why can't the lowest noise or output voltage (-36db) be set to 0000 0000 0000 0001 and the highest noise (0db) to 1111 1111 1111 1111?
[4] I apologize in advance for posting to such an old forum post. I yearn for a greater understanding of this subject.
to add the noob view to what the others said.This article makes it sound to me that when using 24 bit audio instead of 16 bit, the range of loudness that the audio file can describe becomes greater. Why couldn't it instead increase the number of digital increments describing the difference between say -4db and -3db?
additionally, why couldn't the dynamic range of 16 bit audio be expanded by reducing the number of digital increments per 1db change?
Is the difference in db between the integer value that the 16 or 24 bits convert to a fixed value?
To me it seems wasteful to allow a 16 bit audio file to have a noise floor ~60db below the quietest noise in the music (assuming a song with 36db of dynamic range). Why can't the lowest noise or output voltage (-36db) be set to 0000 0000 0000 0001 and the highest noise (0db) to 1111 1111 1111 1111?
I apologize in advance for posting to such an old forum post. I yearn for a greater understanding of this subject.
Thanks in advance!
µ-law/A-law are not "perceptually coded" at all, just companded, which is why they don't work very well for high quality audio.
to add the noob view to what the others said.
PCM is cool because for a given sample, you can easily correlate with an output voltage for the analog signal coming out of the DAC. let's say the first bit can be 0 or 1V, the second bit would code for 0 or 0.5V(half the voltage, giving us -6dB). the third bit codes for half the previous value, and so on.
so one way to look at this makes you and I realize that more bits only let us code quieter and quieter signals. when we add more bits beyond 16, we're pretty much making the background noise having a much better resolution ^_^. that doesn't seem like the best use of bits yet in some ways it is.
Now I do understand why 24 bit can, and always will encode a larger dynamic range than 16 bit, but I'm not so clear on why the added resolution does not mean higher quality. I get that with dither 16 bit can record an essentially perfect waveform with some noise, but wouldn't 24 bit with dither still be better than 16 bit with dither?
to add the noob view to what the others said.
PCM is cool because for a given sample, you can easily correlate with an output voltage for the analog signal coming out of the DAC. let's say the first bit can be 0 or 1V, the second bit would code for 0 or 0.5V(half the voltage, giving us -6dB). the third bit codes for half the previous value, and so on.
so one way to look at this makes you and I realize that more bits only let us code quieter and quieter signals. when we add more bits beyond 16, we're pretty much making the background noise having a much better resolution ^_^. that doesn't seem like the best use of bits yet in some ways it is.
here are other ways to look at this and draw different conclusions:
for example, if we wish to improve the resolution of the musical content you can take my example and think about expressing the signal with 3 bits, then look at what more bits would do(I'm not bothering with how we'd need to code for negative values too on a sine wave, just so that the example is dead simple).
let's say we want the code for 0.4V. with 3 bits 0.4V could be coded using only the second bit to get 0.5V. that approximation isn't great. to increase the resolution between those first bits like you're suggesting, is actually what extra bits do to PCM. when you add an extra bit you can now code for 1V, 0.5V, 0.25V, and 0.125V and turn any combination ON anytime. now we can express our 0.4V as 0.25+0.125=0.375 instead of getting 0.5V with our 3bit code. and the more bits, the easier it becomes to zero in on the desired value.
so in practice extra bits are already increasing the resolution within the audio content like you suggest to do.
another way to look at it is with the additive properties of waves. we can visualize the perfect audio signal, and say that every variation from it is simply another sound being added to the perfect one. the obvious idea for this would be one signal as music and any variation from that being noise. if you increase the resolution so that you can code for a quieter noise, the result is going to look closer to the perfect music signal. it's like turning down the second source of sound. what's left is a cleaner first source. in that respect, just having a low enough quantization noise is giving the music higher resolution.
even if we only cared about not hearing that noise, we would probably wish to keep maybe 11 or 12bit. something like 36dB(about 6bits) wouldn't do for audibly clean background. you want to be sure that the quantization noise(from the value of the lower bit) is so far below the music that we won't notice it. including when the music is not stuck at 0dB all the time. all in all when you start considering pretty extreme examples, and using replay gain or other digital attenuation like changing the volume on the computer, if you still wish for the quantification noise to go unnoticed, you probably won't end up too far from 16bit in a PCM system.
now what I talked about is the correlation between PCM and voltage amplitude of the signal coming out of the DAC. because before and after do correlate, so we can still talk as if a DAC was an perfect R2R design, and get the right results. but in truth most DACs nowadays don't work that way and instead change the amplitude a great many times between each sample. so in practice modern DACs are not using the PCM coding of the file, not as is anyway.
No, 16bit is already perfect there is no "better" than perfect. In fact 1bit is perfect, just with a lot more noise.
Here’s wha perceptual codinhg is:You just NEED to discredit people all the time
Not the way for example mp3s are, but the coding allows increased perceptual quality for speech for which it was developped. The dynamic range of speech exceeds linear 8 bit and µ-law/A-law coding allow bigger dynamic range to be fit to 8 bit at the expence of increased quantization noise which is less harmful than lack of dynamic range. Hence better perceptual sound quality = perceptual coding of a kind even if totally different from perceptual coding methods developped for music/high quality audio.
You know I am right with this so just let it be, ok?
Recording amplitude in one bit would allow you to know if signal was present or not but I don't see how it could do any more than that, and certainly not result in a recognizable copy of the original modulated sine wave. What am I missing?
Lets use the 2 bit audio, we have 11=1.5V, 10=1V, 01=0.5V, and 00=0V. My thoughts were, why couldn't you code your audio file so that... wait, is audio file where audiophile came from???.... code your audio file so that 00 is 0.2V instead. That is what I was sort of thinking.
the obvious example of what he means is DSD. you code in 1bit but you can achieve an equivalent to 24/96 or even higher without too much difficulty. it's one of the best examples to show that whatever we have it's in the end the right accurate signal, and some noise. when we have a way to push the noise around, what's left is the proper accurate signal.I think I understand the application of Nyquist as it relates to sampling rate, but I'm having a little difficulty understanding the '1 bit' statement as it relates to bit depth. Recording amplitude in one bit would allow you to know if signal was present or not but I don't see how it could do any more than that, and certainly not result in a recognizable copy of the original modulated sine wave. What am I missing?
Here’s wha perceptual codinhg is:
https://en.m.wikipedia.org/wiki/Perceptual_audio_coder
µ-law/A-law Is not based on perception at all. So I guess I don’t know you’re right about this after all.
I’m not discrediting anyone, I’m adding accurate information.