Greetings to All,

First, many thanks to Jörn, Bo-Erik, and Michael for your responses to my 
recent post. Your responses gave me food for thought, and I’d like to add a few 
comments regarding audio interfaces, psychoacoustics, and pseudo second-order 
miking (I’m confused in this latter area).

Bo-Erik, I fully appreciate your input. The selection of a good sub is 
something I need to give careful consideration to. I currently use a single sub 
that uses a servo-controlled 12-inch driver in a sealed enclosure. The sealed 
enclosure is preferred because airflow noise through ports is not an issue. 
It’s also a forward facing configuration, so the *directional* characteristics 
of the sub (if direction actually exists) are predictable; or, at very least, I 
can point the driver in a known direction. Unfortunately, this particular sub 
is no
 longer manufactured, but I believe there are a number of decent 10-inch subs 
on the market (and a plethora of junk speakers).

Regarding localization, home theatre sound and the 80-Hz xover point:

I’ll confess ignorance when it comes to knowledge of a separate or unique 
physiological mechanism used to localize (or omni-ize) ultra-low 
frequencies. Within the context of room reflections, music listening, 
home theatre, and the like, I’m fully aware that frequencies below 80 Hz
 are near impossible to localize. Add to the overall auditory scene the 
constituent frequencies that provide unambiguous sound-source 
information, the need for a surround of subs really goes out the window. Some 
spout a slightly higher cut-off frequency, but I gather that 80 
Hz is the accepted standard for home theatre. I have seen literature on 
5.1 and 7.1 referring to subs and low-frequency enhancement for the sole 
purpose of 
effects--the surround speakers are still operated full-range. [Although I 
haven't tried this, I imagine we can accurately lateralize a sub-80 Hz tone 
under earphones. If so, then a lot of our (in)ability to localize low 
frequencies in the sound field is mostly a consequence of physical variables 
such as long wavelengths, head diffraction, room reflections, etc., and not a 
unique mechanism or deficiency of the brain, mid-brain, or peripheral sensory 
organ.] Regardless of accepted protocol, I do have reason for using multiple 
subwoofers, and this reason purposely ignores psychoacoustics.

Although this seems obvious, I’ll argue that sounds in nature do not morph to 
accommodate human perception; i.e., psychoacoustics is a product of a head 
(with ears and brain) in a physical space. For this reason, I do not want to 
present stimuli with “built-in” psychoacoustic enhancements or manipulation. 
Ideally, I want an “untainted” physical space for observing (yes, observing, 
not necessarily measuring) the perception of special-needs populations. 
Populations to be considered include the
 elderly, children who are neurotypical (NT) or autistic and, of course, those 
lucky normal listeners. There's reason to believe that certain populations may 
have compromised auditory processing ability (in addition other sensory input) 
that ultimately results in aberrant or atypical behavior. Because we do not 
know how such persons perceive and react to sound, the best I can do is provide 
physical realism. I wish to present accurate sound-source direction regardless 
of frequency. Psychoacousticians may or may not be in agreement as to how the 
*average* listener perceives sound, but I'm not interested in average 
listeners. Admittedly, some assumptions and subjective impressions always come 
into
 play 
when choosing and using audio equipment. For example, I’ve yet to hear 
any two brands (or models within a brand) of loudspeakers that sound 
anything alike, so we have to accept that none of this is
 going to be perfect.

When it comes time to construct a sound system for accessing sound quality (or 
simply for musical enjoyment), I will most certainly use a single sub as you 
suggested. But for my proposed system, the subwoofer's crossover frequency and 
filter order does become something of a choice based on speaker performance. 
Because I will be filtering/processing the four B-formatted wave files before 
decoding (none of the processing will be done in real-time), I have a lot of 
choices for filter types--and perhaps the addition of group delay. I have 
numerous MATLAB Toolboxes for processing wav files in addition to the Advanced 
Signal Processing and Digital Filter Design Toolkits in LabVIEW. Thankfully, 
I’m no longer limited to the *bouncy* 8th-order elliptic filters I used to 
construct. The problem nowadays is that there are way too many choices that are 
relatively easy to implement. Responses to my last post provide clearer 
direction--thanks.

Jörn brought into the discussion not only filter type and slope, but the choice 
of audio interfaces. I have first-generation MOTU 896HD units. Although there 
are two FireWire ports on each unit's backside, MOTU states that the only two 
units can be synchronized using a FireWire (daisy chain connection), and the 
combination of more than two interfaces requires a
 master clock (the units have word clock in and out). There is an optical 
interface in addition to the AES Pro (XLR). My gear is currently out of reach, 
but I’m guessing that the optical connect is intended for ADAT/Lightpipe, not 
optical S/PDIF (they're two distinctly different protocols and not 
interchangeable). A separate 8-channel A-D could certainly be used with ADAT. I 
suppose there’s no reason to worry about inter-channel timing issues when using 
dissimilar components (meaning a MOTU optically linked to an D-A device). 
Similar to the MOTU interface, my M-Audio ProFire 2626 provides a lot of input 
and output options (two ADAT ports), but a D-A converter would still be needed 
for > 8 analog outs.

I like the robustness (and XLR) connects of the MOTU 896; I’ll admit that much 
of this is a personal choice but it's not meant to promote or discount any 
single piece of gear. When it comes to configuring hardware and software, I 
don’t know whether all DAWs provide the option of assigning tracks to all of the
 physically available ports (for example, one of my USB interfaces permits a 
choice of digital OR analog, but not both simultaneously). Furthermore, I want 
the presentation of stimuli to be glitch-free. I imagine most modern high-end 
DAWs and interfaces provide crash-free performance, but mixing 48 mono tracks 
to stereo isn’t the same as providing 48 discrete analog out channels when it 
comes to stable performance. Perhaps my fear of computer crashes (both Mac and 
PC) comes from past experiences. What I'm using now seems glitch- and 
crash-free, hence my desire to stick with it (and the 16-channel count).

Regarding speaker arrays: Thanks, Jörn, for suggesting a large (ear-level) ring 
with smaller rings above and below the larger ring. My original idea (two large 
rings) stemmed from the notion that the 12 speakers would lie on the surface of 
a large (virtual) sphere whose poles would extend beyond the room dimensions, 
thus giving the impression of a *bigger* listening space. Of
 course, a sense of distance and spaciousness is intrinsic to the recording, 
not how far the actual speakers are from the listener (well, we could get into 
a wave-front curvature discussion, but I’m not ready for that). The other 
reason had to do with placement of video monitors. If the video doesn’t 
interfere with the speaker array, I’ll make drawings for the speaker layout you 
suggested. Thanks. I may have access to a 10 x 20 foot room. I don't know the 
ceiling height. To my knowledge, all surfaces are treated with 6- or 8-inch 
foam. From what I've been told, the room is practically anechoic down to 200 
Hz. Maybe the leftover space could be used for bass traps, aborbers, or 
diffusers to further tame the low frequencies. I might have access to a B&K 
intensity probe for calibration. It would be interesting to look at the 
velocity and pressure components resulting from the surround system as measured 
from the listening position. This could give some
 measure of wave field accuracy that goes far beyond perceptual judgments. 

RE mics: The idea of going *second-order* occurred to me, but I’m too ignorant 
on this topic to say much. I’ve overheard discussions (mostly at trade shows) 
where first-order mics could be *stretched* to give second-order performance 
(and further stretched to give HOA performance or approximations). As I 
understand, Ambisonics can be enhanced by the addition of U and V components, 
and these are extensions of the B-format components derived from a first-order 
mic (based on the raw, or A-format, data). From what I've read, U and V 
contribute to horizontal, not
 vertical, image stability. With regard to live recordings, only a HOA-specific 
mic (e.g., VisiSonics or Eigenmic) can provide true HOA performance. Adding 
more playback channels to recordings obtained with a first-order mic may give 
better room coverage/distribution, but nothing is gained in terms of accurate 
*wave-field* reconstruction. Am I correct here? Although I could create 
higher-order  stimuli via modeling, the plan is to use complex and dynamic 
stimuli (that includes moving objects such as automobile traffic) obtained via 
live recordings. The emphasis is on *real-world* reconstruction of 
representative environments, not listening assessments based on movie-goers 
expectations (movie sound designers provide us with ample conditioning; heck, 
you can even hear sounds in the vacuum of deep space!).

As always, many thanks for your time and input.
Best,
Eric C. (by the way, I started putting C here so that readers don’t confuse me 
with the Erics who are
 10X smarter than yours truly).
-------------- next part --------------
An HTML attachment was scrubbed...
URL: 
<https://mail.music.vt.edu/mailman/private/sursound/attachments/20130221/c45ed548/attachment.html>
_______________________________________________
Sursound mailing list
Sursound@music.vt.edu
https://mail.music.vt.edu/mailman/listinfo/sursound

Reply via email to