johncarm
100+ Head-Fier
- Joined
- Nov 10, 2013
- Posts
- 486
- Likes
- 82
In asking the question, "What sample rate is needed for digital to be perfect," we often refer to the ear's upper bandwidth. If the Nyquist frequency is greater than the ear's bandwidth, we're good.
That might be a bad assumption, however.
I've been reading through some sound science posts here and there's an amazingly pervasive assumption in the threads here about how the brain works, which I can't imagine could possibly be true, although I would like to see what psychoacoustics or neurology says about it.
Here's the basic idea. Start with two signals A and B, and we would like to see if the brain can tell them apart. I'm not talking about "proving" anything "for all humans" etc., just some basic idea of what we need to know to set up some tests, or make some reasonable guesses.
In scenario one, A is some white noise at 0 dBFS. Let's say we have a recording of a sine wave at -10 dBFS, called M. We generate B by mixing A and M.
Here, we need psychoacoutical facts regarding the ability to hear smaller signals in the presence of larger ones. I'll call this "masking theory" for short. Both A and M are simple signals by comparison to music, and even if masking experiments were done with sine waves or noise bursts, they would be relevant.
In scenario two, A is a choir, M is a marching band at -60 dBFS, and B is the mixture of A and M. Now the signals are more complicated, but they share a feature with scenario one, which is that A and M are highly differentiated. Either one can be heard as a coherent, recognizable sound source. And, when mixed, they will "clash" rather than blend into a single apparent phenomenon.
In both scenarios one and two, it makes sense to view signal B as "A plus something added" and to examine the features of the signal that was added. The subtracted signal B-A is something that makes sense to the brain in its own right.
In scenario three, A and B are the output of two different amplifiers, both fed with the same input. Now this is quite different, because there simply is no meaning to B-A. The subtraction is hardly relevant to how the ear hears A and B. This means that psychoacoustical research on "small signals in the presence of larger ones" would have to be applied selectively, and only then with a good justification (such as a model of the ear that justifies it).
In scenario four, A is some music that is recorded with a bandwidth of 100 KHz, but then band limited to 20 KHz, and B is the same recording but band limited to 40 KHz.
This is where it's tempting to say "B-A is entirely ultrasonic," therefore the difference can't be audible. But that is conflating two different questions. The first question is "Can I hear a pure sine wave in the range 20 KHz to 40 KHz?" to which the answer is probably "no." The second question is "Can I tell A and B apart when it happens that the difference between them contains only components above 20 KHz?" That is a different question with possibly a different answer. That is the real question to ask about higher bit rates.
That might be a bad assumption, however.
I've been reading through some sound science posts here and there's an amazingly pervasive assumption in the threads here about how the brain works, which I can't imagine could possibly be true, although I would like to see what psychoacoustics or neurology says about it.
Here's the basic idea. Start with two signals A and B, and we would like to see if the brain can tell them apart. I'm not talking about "proving" anything "for all humans" etc., just some basic idea of what we need to know to set up some tests, or make some reasonable guesses.
In scenario one, A is some white noise at 0 dBFS. Let's say we have a recording of a sine wave at -10 dBFS, called M. We generate B by mixing A and M.
Here, we need psychoacoutical facts regarding the ability to hear smaller signals in the presence of larger ones. I'll call this "masking theory" for short. Both A and M are simple signals by comparison to music, and even if masking experiments were done with sine waves or noise bursts, they would be relevant.
In scenario two, A is a choir, M is a marching band at -60 dBFS, and B is the mixture of A and M. Now the signals are more complicated, but they share a feature with scenario one, which is that A and M are highly differentiated. Either one can be heard as a coherent, recognizable sound source. And, when mixed, they will "clash" rather than blend into a single apparent phenomenon.
In both scenarios one and two, it makes sense to view signal B as "A plus something added" and to examine the features of the signal that was added. The subtracted signal B-A is something that makes sense to the brain in its own right.
In scenario three, A and B are the output of two different amplifiers, both fed with the same input. Now this is quite different, because there simply is no meaning to B-A. The subtraction is hardly relevant to how the ear hears A and B. This means that psychoacoustical research on "small signals in the presence of larger ones" would have to be applied selectively, and only then with a good justification (such as a model of the ear that justifies it).
In scenario four, A is some music that is recorded with a bandwidth of 100 KHz, but then band limited to 20 KHz, and B is the same recording but band limited to 40 KHz.
This is where it's tempting to say "B-A is entirely ultrasonic," therefore the difference can't be audible. But that is conflating two different questions. The first question is "Can I hear a pure sine wave in the range 20 KHz to 40 KHz?" to which the answer is probably "no." The second question is "Can I tell A and B apart when it happens that the difference between them contains only components above 20 KHz?" That is a different question with possibly a different answer. That is the real question to ask about higher bit rates.