If you can use DSP to convolute HRTF-impulse responses with the music then of course do that. That's cross-feed too, just in more sophisticated form.
Every time you and pinnahertz exchange ideas, we all learn. That’s dialectics.
I have some doubts.
Correct me if I am wrong, but you said that convolution of HRTF-impulse responses is a more sophisticated form of crossfeed.
I was wondering that I would be able to measure with two microphones in my ears a binaural room impulse response. Then I thought such BRIR would “inseparably” contain not only my HRTF but also early reflections and reverberation from the room I measured in. But then I could possible (but way less probably) measure impulses inside an anechoic chamber and then measure only my HRTF.
So when you convolve a BRIR I see the reason for adding electronic crosstalk, because you are trying to replicate the whole system composed by the rig and room itself.
But when and why you would want to add electronic crossfeed if you convolve a generic HRTF or you had the opportunity to measure your own pure HRTF?
Would it be anyhow useful to avoid acoustic crosstalk with speakers playback or to avoid electronic crossfeed when you convolve even a BRIR with headphones playback?
What is the difference between a) the computation of the localization of sound object in a separate tracks (within a digital stream of many sound objects tracks) using a HRTF with density of lets say 720 measured coordinates (360/5 azimuth locations * 10 elevations) and b) the interpolation* of two coordinates (+ and - 5 degrees azimuth, zero elevation) between a 360 degrees of head movement freedom when playing binaural content**?
* interpolation for head tracking purposes
** or content with natural ILD and ITD; I see you have a “DIY Jecklin disk microphone” in your profile...
Would you need electronic crosstalk in such headphone playback environments with such content (binaural recordings)?
What happens if a generic or your personalized HRTF does not have a 720 density but let’s say 16 coordinates density and you need interpolation to calculate sound objects in their designated locations? The function to do that interpolation is the same function would you use to head tracking?
So if you have a HRTF of 16 coordinates and you apply interpolation not only to compute the sound object location but to rotate then acording to the user head would you still need crossfeed when playing Atmos content? And if you are playing third order ambisonics content?
If you want my opinion, I guess Atmos is intrinsically limited to and ambisonics was already designed to deal with acoustic crosstalk, but I am curious to know what would happen if you do not add electronic crossfeed*** when convolving a HRTF with headphones playback of Atmos and higher order ambisonics content.
*** Or deliberately and carefully control its level in different and lower intensities we would find in acoustical crosstalk with speakers; when dealing with acoustical crosstalk I am referring to speakers located on both hemisphere that are cut by the median plane. The speakers in the same hemisphere would be summed with the HRTF filter anyway and the speakers placed within the median plane would also be filtered by the HRTF without any electronic crosstalk setting.
Any idea?