Hello Etienne and all Sursound Readers,

Many thanks for your response, insight, and “food for thought”. You brought up 
interesting points which, in turn, prompted me to dig deeper into Ecological 
Psychology (referring to the Gibsonian school).

There’s certainly something to be said for choosing the “right” information 
versus ability to detect or pick up additional information. In fact, this could 
get to the heart of some of my initial thoughts regarding hearing research (and 
my initial interest in Ambisonics). As you may know from prior diatribes and 
posts, I have interests in cochlear implant research and spatial hearing. It 
probably comes as zero surprise that an array of 22 electrodes used to 
innervate the auditory nerve provides, at best, impoverished input to the brain 
(especially when compared to the
 input provided by the approx. 3500 inner hair cells of normal-hearing 
listeners).

When electric hearing is combined with acoustic hearing (hearing aid or not), 
we might surmise that the low-frequency (acoustic) energy simply adds to the 
amount of information received. For normal-hearing and impaired listeners 
alike, the low-frequency energy by itself provides very little usable 
information. For example, low-pass filtered speech (f3 = 200 Hz, high-order 
filter) is quite difficult to understand. In fact, f0 for women is above 200 
Hz, so little speech information resides at the very low lows.

Electric (cochlear implant) hearing alone provides reasonably good speech 
comprehension scores when speech stimuli are presented in a quiet environment 
(+20 dB or better SNR). Scores obtained from 5-word sentences could range from 
50 – 90 percent correct (I don’t have an exact reference at hand, but I believe 
this is a good estimate). When electric
 hearing is augmented with the below-200 Hz acoustic stimulus, speech scores 
improve by a big jump. Furthermore, speech comprehension ability in reverberant 
environments improves. One might be inclined to conclude that when the sensory 
input is impoverished, any additional input is welcomed and quickly used to 
fill in any missing information gaps or resolve ambiguities. But the 
synergistic combination of electric and acoustic hearing suggests, at least to 
me, something beyond “additional” information is at work.

Research regarding electric-acoustic stimulation (EAS) has led to exciting 
results and interesting discussions, but the background noise and reverbation 
used in many studies are often of the artificially-generated (pink or white 
noise maskers) and one-dimensional or mono reverb nature. Sursound Digest 
readers probably recall the discussion I initiated regarding multi-channel 
subwoofers and identifying sound-source direction (at least
 in free-field) for very low frequency sounds. My interest and concern for 
presenting accurate low-frequency and realistic sound source direction wasn’t 
about measuring localization ability for very low-frequency sounds: My interest 
was to build a periphonic system for evaluating REAL-WORLD low-frequency 
sounds’ contribution to or detraction from EAS listening. Needless to say, 
real-world sounds don’t come from a single subwoofer or direction. Whether we 
can determine direction isn’t the important part, but the subtle (and perhaps 
subconscious) aspects of real-world listening do matter.

My take or concern over “realism” versus artificially produced stimuli isn’t 
one of difficulty; in fact, I’d state that many artificial and monaural noises 
(dichotic or diotic presentation) mixed with speech present more difficult 
listening conditions than what we encounter in the real world. The problem is 
one of learning what is “real” and useful. As an analogy, being able to ride a 
unicycle (arguably difficult) doesn’t guarantee one’s success or ability to 
ride a bicycle. It may be easier to ride a bike, but there are also more ways 
to fall or crash at high speed, so the need to maneuver a bike and learn the 
rules of the road are more important for safety. Learning, success, or 
attending to a difficult listening task in the laboratory doesn’t guarantee 
user success in the cacophony of real-world stimuli.

To date, I’m not aware of studies where real-world stimuli and scenarios have 
been used to study the efficacy of EAS listening. It is entirely possible that 
the addition of low-frequency energy, whether a part of speech signal or not, 
helps users choose the “right” information. In the real world, there is a lot 
of multi-sensory noise. For
 normal-hearing listeners, segregating an auditory signal from noise is 
accomplished, in part, by perceived spatial separation. This is equally 
important for those involved with the recording arts: Spatially separating a 
pair of shakers from a hi-hat is most often accomplished via panning. With 
speech, we also face informational masking as well as energy masking. So, 
adding more “speech” information (aside from level to improve SNR) isn’t as 
important as our ability to choose information deemed important to the task at 
hand. We do learn from experience what is “true” or “correct” and is why we can 
learn to localize with fake pinnae or even develop localization ability with 
ONE ear (animal research supports this). Cochlear implant recipients need time, 
too, to make use of their “new” sense of hearing.

Another area I would be interested in investigating is time-to-contact as it 
applies to hearing (Gibson was mostly involved with
 vision), and how binaural implantation might improve a listeners sense of 
safety in “three dimensional” space where there are multiple, moving, sound 
sources. Such studies under headphones are very realistic. As you wrote, “One 
of the characteristics in Gibson’s ecological approach that has been adopted by 
the VR field is the idea that perceptions are confirmed as true through 
‘successful action in the environment’. Tilting one’s head can be considered 
action in the environment, and if the spatiality of the sounds heard correlate 
then that action can be considered successful. So head movements help to 
confirm that what is being perceived is correct.” I very much agree with what 
you wrote. Adding to this, avoiding collision is certainly a more successful 
action than identifying location to the nearest nth of a degree.

Thanks for taking time to read. I always look forward to the Sursound Digest 
regardless of topic.
Best always,
Eric C.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: 
<https://mail.music.vt.edu/mailman/private/sursound/attachments/20130402/9e04d7fd/attachment.html>
_______________________________________________
Sursound mailing list
Sursound@music.vt.edu
https://mail.music.vt.edu/mailman/listinfo/sursound

Reply via email to