[1] I think you are interpreting clarity as "separation" or "the ability to hear individual instruments." I mean something else.
[2] At one moment in time, there is a set of instruments playing. Usually the musicians (and composer) intend for some of those to be foreground, and some background. It may be that the presentation is supposed to have at most one foreground sound, or two. There may be a kind of hierarchical structure, in which sounds go progressively into the background. The structure may be about note attacks and timing of attacks as well as the phenomena of sustained sounds.
[3] If you hear every single violin, the result is not clear. That will not make the hierarchical structure of foreground and background clear. It will sound muddy as a whole. I have some chamber music recordings in which the individual instruments do not blend at all.
[4] I also mean clarity of patterns over time. For instance, the musicians may shape the dynamics of a phrase. And that shape relates to the shape of previous and successive phrases. The question is how clear this shape, and the relationship of shapes, is.
1. I think you must be explaining poorly what you mean by clarity or misunderstood what I stated, if you really "mean something else". Because ...
2. How is separating (groups, sub-groups or individual) musicians into "foreground" and "background" not "separation"? Would you generally not want/expect greater clarity of those musical entities which are supposed to be in the foreground or put differently, a level of clarity which allows for a differentiation/separation between foreground and background?
3. Notice I've used the invented term "musical entities", a musical entity might be a single musician or might be a group/sub-group of musicians, depending on the composition (orchestration) at any particular point. Regardless of who actually comprises a musical entity at any one point in time, in the case of one or more musical entities in the foreground, we would want/expect greater clarity of that foreground entity (relative to other entities in the background) but not of the individual musicians who comprise that entity, unless of course the entity is a single musician. Your statement appears to entirely agree with what I'm saying! So I'm not sure why you appear to be arguing?
4. Again, this is effectively the same thing I'm saying!
BTW, it's dangerous/potentially misleading to use the terms "foreground" and "background". I know that you are talking from a musical perspective and essentially mean "to the fore" but we need to be careful "foreground" and "background" also have specific meanings in terms of physical geography, which we can also represent in audio recording/reproduction. We do this almost constantly in TV/film audio, commonly in popular music genres but less so in orchestral music, except maybe in the case of those few pieces which employ say off-stage brass, in which case we maybe looking at the off-stage brass as being both background geographically but foreground musically.
[1] You could argue that science should not be investigating art because "art is subjective" but I disagree. I'm sure we agree that science should be investigating perception, but if you express an interest in perception, [2] I don't see how you can fail to acknowledge that there's a very important and universal phenomenon called art.
1. I'm not arguing that science shouldn't investigate art. However, science hasn't got particularly far in this regard, it's still trying to work out many of the basic individual aspects of perception, let alone how all the aspects of perception combine/integrate to into an appreciation/evaluation of art.
2. This statement appears deliberately disingenuous just to prove a point. The term "art" exists universally (AFAIK) in all languages but there is no universal phenomenon called art. Certainly there are some individual works which would most likely be universally accepted as art, the Mona Lisa or Beethoven's 5th for example but there is no universal definition of what art is. While it's an interesting philosophical question, to investigate it scientifically, we have to know what "it" is and not only is there no universal agreement of what it is but we're not even close to a universal agreement, opinions can vary diametrically and occasionally, even to the point of violence!
You say "perception is different from reality" but that's not the right way to look at the problem. There is the reality of how perception works. That's a reality. And it affects how we perceive music and recordings.
I'm not sure I understand what you're saying, what is the right way to look at the problem? If I see a pig (A) and imagine that pig flying (B), I can accept that my imagination itself exists in reality but not that what I'm imagining does, I don't accept that flying pigs actually exist. If I want to understand something about the process of imagination, isn't comparing the difference between A and B a good place to start and then coming up with theories/experiments to explain the difference? What's the alternative, coming up with theories/experiments to explain how pigs fly?
[2] I think my central question is--why choose one mic position over another? [1] I think the best reason for choosing one mic position over another is how it affects the musical patterns as heard in the control room, and that means the engineer needs to be hearing those patterns in the first place (and care about those patterns). [3] Judging which of two mic positions gets closer to the original musical pattern is at least 50% a musical decision.
1. You would think that, you are a musician! A musician is defined by their musicality and trains their hearing/perception to be sensitive to the evaluation of musicality. A recording engineers job is to record an acoustic signal as best as possible and musicality is largely irrelevant. If I perfectly record a terrible musical performance, I've done my job fabulously well! As a recording engineer I'm primarily listening for: mic frequency and amplitude response, phase artefacts between mics/input signals, signal to noise ratios of the inputs, various other potential interference/unwanted artefacts, the relative amplitude and frequency response (balance) of direct vs indirect acoustic signals being picked-up by the individual mics and combination of mics and, by extension of these factors, stereo imaging. These aspects of recorded sound are in my control and are my responsibility as a recording engineer, musicality isn't! Obviously, a great recording of a terrible performance or a terrible recording of a great performance are both undesirable results and that is why the role of Producer exists. This is one of the missing "holes" in your assumption which I mentioned previously. The producer is listening for musicality and how that musicality translates (through the speakers) and in practise is a sort of bridge/arbiter between the musician and the engineer. Ignoring some of the practicalities/technicalities of the role, a musician could potentially take on the role of producer relatively quickly, not so much the role of engineer though!
2. To best satisfy those things I'm listening for, mentioned above. It was discussed previously whether a musician has any idea of what the audience hears during a performance. I would say they do, although only a very generalised vague idea or at least very generalised and vague compared to the recording engineer. A musician will tend to think in terms of "the audience" and of how an acoustic affects the audience perception, because the musician only has some degree of control at that level. The recording engineer has a great deal more control and therefore thinks in terms of individual audience members or sections of the audience and the different acoustics in different parts of the venue. The goal of the engineer (and producer) would commonly be some sort of mean or biased mean of these difference acoustical positions. That commonly results in an illusion of an acoustic space rather than a reality, in the same way that no family actually has 2.4 children, even though that may be the average family size.
3. Here we enter murky ground. I've painted a very black and white picture above of the roles of engineer, producer and musician in order to more succinctly explain them but in reality, it's usually rather more grey. In practise, the engineer will design a mic'ing scheme, depending on the acoustics of the venue, to fulfil both the fundamental engineering requirements and the desire/s of the producer. Mic choice/positions therefore does in practise include musical perception considerations and a good, experienced recording engineer will have picked-up a considerable amount of that over the years. Typically IME, particularly when working with an new producer, there would be a meeting well before the first recording session, so the engineer can gain an insight into the desires of the producer and therefore design a mic'ing scheme likely to facilitate those desires.
I also am suggesting that fidelity is a musical observation. To give an extreme example to make a point, an engineer trained only in sound fields will not be in a position to evaluate fidelity.
Doesn't this bring us back to what I mentioned above? What is fidelity in relation to, is it the relationship/similarity between two real/actual signals (say in the case of digital audio data input into a DAC and the resultant electrical signal it outputs) or is it the relationship/similarity between a perceived experience of a live performance and the audio recording/reproduction of it? In the case of the former, we can turn around your argument because a musician (or any human being) is not in a position to evaluate fidelity because no human can hear digital data or an electrical signal or compare them, humans can only hear acoustic signals. The case of the latter introduces interesting factors: Of course we cannot reproduce a perceived experience of a live performance with an audio recording because an audio recording only contains audio, not any of the other factors which contribute to a perceived experience. Having said this, there are some generalised aspects of perception which we can compensate for, by "compensate for" I mean change the reality to help create a better representation of the perception. For example, in the mid/late '80's, a few of the more cutting edge labels were taking advantage of digital audio and in addition to single position mic arrays they added spot mics (mics positioned to pick up individual musicians or a small group of musicians within the orchestra). This allowed the spot mic output to be added to the mic array output in appropriate places, to emphasize that instrument, to a similar subjective level as would occur in a live situation when a visual cue would cause a perceptual emphasis. The use of this perceptual effect is far more prolific in the film sound world, where it's known by the term "hyper-reality" and has been used as an audience manipulation tool for over 6 decades.
All the above relates mainly to orchestral music, from the 1960's popular genres of music evolved to take advantage of recording technology and the concept of realism is therefore a largely abstract one.
G