Greetings All,
I continue to learn a lot from this site (and someday hope to have something to 
give back). For now, however, I will briefly comment on posts by Umashankar 
Mantravadi, Augustine Leudar, and Entienne.

Entienne wrote the following [abridged]: **The argument essentially says that 
for something to appear real it has to fit people's *pre-conception* of what is 
real, rather than fit what actually is real. In other words, throw out 
veridicality (coincidence with reality); instead, try to satisfy people's 
belief of reality. This is another argument for questioning the extent to which 
physical modeling has the capacity to create illusions of reality in sound...**

Entienne made me consider further something of great importance re my CI 
research. Briefly, we really don’t know what auditory perception is like for 
hearing-impaired listeners (remembering that there’s a lot more to 
sensorineural hearing loss than threshold elevation). For example, does the 
Haas effect work for them? Why is noise-source segregation so difficult? Does 
breaking apart an auditory scene create greater dysfunction, or can they put 
the pieces back together to give the illusion of a unified sound source (as 
with the cello example)? How does multi-band compression sound for them, etc? 
We would most certainly like to know how altering a physical stimulus improves 
their belief of reality (thus improving their ability to communicate or enjoy 
music)? But how do we measure the perception of cochlear implant and hearing 
aid users other than providing *physically accurate, real-world* stimuli? Side 
note: Thanks for the reference to H. Wallach
 (1940).

Re Augustine’s post: Thanks for suggesting Gary Kendall’s paper. While it 
doesn’t provide a *complete* explanation (who can?), it is a good read. I 
proposed a somewhat similar study while a grad student, but the stimuli would 
have included speech, dynamical sounds (such as breaking glass or a bouncing 
ball), and unfamiliar sounds. The constituent components of the unfamiliar 
sounds would be spatially separated but have identical start times. We could 
then ask whether it’s familiarity (as with a cello), arrival times, or other 
variables that unify the separate sounds into a common source.

Umashankar Mantravadi wrote the following: *As a location sound mixer, I 
exploited the visual reinforcement of sound in many situations. If you are 
recording half a dozen people speaking, and the camera focuses on one - 
provided the sound is in synch - the person in picture will sound louder, 
nearer the mic, than the others. It is a surprisingly strong effect, and one 
side benefit is you can check for synch very quickly using it.*

Many thanks for sharing this experience. I am currently creating AV stimuli 
(using a PowerPoint presentation as the metronome/teleprompter). While there is 
nothing new or novel about incorporating video, I am unaware of any 
investigations using cochlear implant patients’ in a surround of uncorrelated 
background noise combined with a video of the talker(s). One could also study 
the effects of simulated cochlear implant hearing (using normal-hearing 
subjects) with visual cues in a *natural* environment.

It has been known for some time that lipreading is useful for comprehending 
speech presented in a background of noise. For example, Sumby & Pollack (1954) 
showed that visual cues could aid speech comprehension to the same degree as a 
15 dB improvement in SNR. Sounds with acoustic features that are easily masked 
by white noise (for example, the voiceless consonants /k/ and /p/) are easy to 
discriminate visually.

There is a plethora of literature surrounding the benefits of lipreading. It is 
entirely possible that visual cues can affect more than just speech 
comprehension. A study showing the reduction of stress when a listener is aided 
by lipreading could be interesting: It is possible that visual cues, regardless 
of speech comprehension advantages, could reduce listener stress in a difficult 
listening environment. Capturing subtleties, such as talker voice level as a 
function of background noise level, could make video and audio stimuli more 
realistic. Although we might not be sensitive to these subtleties when 
sufficient information is available to us (us = normal-hearing listeners), 
hearing-impaired individuals might make use of visual cues in ways that have 
not been explored. Systematic reduction or elimination of available information 
can be accomplished when the stimulus contains all of the *essential* 
information in the environment.

I am exploring Ambisonics (along with video) as method of capturing the 
essential information. At worst, I have a great-sounding system for listening 
to others’ musical recordings. Thanks to everyone for the recordings of crowd 
sounds, music, software, and for sharing your wisdom.
Best regards,
Eric
-------------- next part --------------
An HTML attachment was scrubbed...
URL: 
<https://mail.music.vt.edu/mailman/private/sursound/attachments/20120602/bb7c646d/attachment.html>
_______________________________________________
Sursound mailing list
Sursound@music.vt.edu
https://mail.music.vt.edu/mailman/listinfo/sursound

Reply via email to