Did anyone read this article? I don't mean quote mine to cherry pick support for your existing firmly held views. I mean read the article.
I might just eventually read it for my own interest. But if someone else has read it (or plans to), I'd be interested in discussing whether/how it informs the discussion in this thread.
Debates about what you guess it says are valueless for me.
I now have read a significant part of the article (from the start to a bit into the part headed "Methods"). It sure is very interesting.
To the question whether/how it informs the discussion in this thread: Maybe in some ways but not like it gives a direct and conclusive answer or without some additional kinds of experiments.
The experiments done are with a limited number of sounds that are mostly created in a virtual room (or for some experiments a virtual anechoic chamber), where reflections (up to 0.5s) are added, and then all the resulting sounds arriving at the subjects position (the subject being the "artificial listener") are binauralised using the HRIR of a dummy head with torso.
(Next to the head and torso represented by the HRIR the "Artificial listener" roughly consists of a virtual cochlea and then the actual neural network.)
So in itself it is based on localizing a limited number of sounds in the environment the natural way, wich is something different from localizing sound in a (normal) stereo recording played over speakers (for example).
But you could of course present the virtual listener with a binauralised signal based on two loudspeakers in the virtual room, that are playing a recording with a limited number of sounds panned somewhere between the speakers, and just see what happens! Maybe the algorithm will simply identify 2 sound sources (the 2 loudspeakers) if it is not "susceptible" to the stereophonic illusion. Or maybe it is "susceptible" to the stereophonic illusion and it will identify sound sources in between the speakers.
(But standard panning is just one of many "stereo tricks", so this is just scratching the surface in a way.)
For this experiment one could of course send the stereo recording used through a DAC and ADC before going into the virtual room.
(For the sake of this thread discussion. I don't seriously see how it could affect anything unless the DAC had very serious issues. Let me note here that in the experiments they also added sources of noise in the virtual room, with SNR ranging from 5 to 30 dB, hard to see how DAC imperfections below -90 dB or -100 dB could be a problem if those noise sources were not. And channel synchronisation of DACs should be good enough by a large margin I think.)
Another way to test the influence of different DACs could be by just using one channel of the DAC to generate one of the sounds to be localised in the standard experiments (not concerning the stereophonic illusion). But then the most likely (or rather least unlikely) potential cause of problems by the DAC - any differences in handling the left and right channels - would not play a role.
Anyway, regardless of the relevance to this thread discussion I find the article itself very, very interesting.
The model shows many properties (and limitations) similar to those of human hearing.