[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: Number of concurrent streams.



Dear Claire and list,

I just made up the term "matching a standard" when I sent the
last message.  A discussion of this and other methods of
measuring stream segregation is given in :
Bregman, A. S., (1990/1994) "Auditory scene analysis.", in the
section entitled, "How to measure streaming", pages 52-58.

The method I described in my message is described in the
sub-section "Pattern recognition" on p.56.

Good luck.

Al

----- Original Message -----
From: "Claire Piché" <clairepiche@VIF.COM>
To: <AUDITORY@LISTS.MCGILL.CA>
Sent: Monday, May 03, 2004 8:43 AM
Subject: Re: Number of concurrent streams.


> Dear M. Bregman
>
> Can you tell me more information about the method you called
"matching a standard", Is it a syntagm of your own? Actually, i'm
writing my thesis and a few days ago i started the part where
your work is the main figure. Moreover, the second part of your
letter reaches the heart of my work.
>
>
> Claire Piché
> Doctorat en Études et pratiques des arts
> Université du Québec à Montréal
>
>
>
>
>
>
> Al Bregman a écrit :
>
> >Dear Valeriy and List,
> >
> >(Sorry if you get this twice.  Apparently the AUDITORY
listserv
> >didn't like my new email address and may or may not have sent
> >this message to the list.)
> >
> >I don't think that the question about how many sound sources
> >people can perceive in a mixture has a simple answer.
> >
> >First of all, it depends on what you mean by "perceive".  If
you mean how many sound sources they can report, then it depends
on how much time they are given.  The longer they listen, the
more they can report.  This could mean that they can pay close
attention to only one sound at a time, but can shift their
attention from one source to another.  If you mean, "How many
could they report if given a sample of each one in turn and asked
whether they could detect it in the mixture, given all the time
they needed?", the question becomes one about blending and
segregation at a very basic level. (Let's call this last method,
"matching a standard").  Often I haven't been able to hear a
sound in a mixture until I knew what it was.
> >
> >Assuming that one is using the "matching a standard" method,
> >then one's success will depend on what the sounds are and how
intense they are.  Obviously a weak sound may be hard to
> >detect -- it may be masked (psychoacoustically or
> >informationally) by the others.  Even if this is not true, the
similarity of the component sound sources plays an important
role.  Those researchers who have claimed that only a small
number of sources (3 or 4) can be detected are all referring to
sets of sounds that resemble one another, such as multiple
talkers or singers, or rhythmically playing instruments.
> >
> >Consider the following set of sounds:
> > - a person talking
> > - randomly spaced hits on a bass drum (greatly attenuated)
> > - an ambulance siren
> > - jangling of a set of house keys,
> > - a pure tone playing Morse code,
> > - a person typing on an electric typewriter.
> > - the sound of a motorcycle whizzing by (greatly attenuated)
> >
> >As long as the intensities were balanced appropriately, I
think
> >you would eventually detect all of them by the method of
matching a standard.
> >As many of the list members know, I believe that there is an
> >early perceptual stage of assigning of links among the parts
of the incoming mixture prior to further processing by the
mechanisms that we call attention. If what we are asking about is
whether this pre-attentive process has some limit concerning the
number of discrete subsets (potential streams) it can form, we
would have to observe its operation without any contribution from
attention.  I believe that
> >this is impossible using the standard methods of
psychoacoustics. Rather, it has to be addressed using a
physiological approach. Some beginnings toward doing this have
been carried out by Elyse Sussman and by Claude Alain (working
independently), and there may be others who I don't know about.
> >
> >By the way, I referred to the output of the pre-attentive
> >mechanism as "potential" streams because there is good reason
to believe that top-down processes play a big role in determining
the actually heard streams.
> >
> >Sorry the answer couldn't have been simpler.
> >
> >Al
> >
> >---------------------------------------------
> >Albert S. Bregman,
> >Emeritus Professor
> >Psychology Dept., McGill University
> >1205 Docteur Penfield Avenue
> >Montreal, Quebec
> >Canada  H3A 1B1
> >
> >Office:
> >     Voice: +1 (514) 398-6103
> >     Fax:     +1 (514) 398-4896
> >---------------------------------------------.
> >
> >
> >
> >----- Original Message -----
> >From: "Valeriy Shafiro" <Valeriy_Shafiro@RUSH.EDU>
> >To: <AUDITORY@LISTS.MCGILL.CA>
> >Sent: Friday, April 30, 2004 3:26 PM
> >Subject: Re: Computational ASA -- how many sources can humans
> >perceive?
> >
> >
> >
> >
> >
> >>I would like to ask a further question: Do we, in fact, know
> >>
> >>
> >how many
> >
> >
> >>independent sound sources in a mixture humans can perceive?
> >>
> >>
> >Thus far I
> >
> >
> >>know of only one research report where human listeners were
> >>
> >>
> >asked to
> >
> >
> >>identify sound sources in a recorded "real-world" sound
mixture
> >>
> >>
> >(Ellis, D.
> >
> >
> >>P. (1996). Prediction-driven computational auditory scene
> >>
> >>
> >analysis).  We
> >
> >
> >>have been talking about this issue with Brian Gygi, and from
> >>
> >>
> >the few
> >
> >
> >>related reports that Brian found, it appears that humans may
> >>
> >>
> >not be that
> >
> >
> >>good in simultaneous perceiving independent sound sources.
For
> >>
> >>
> >instance,
> >
> >
> >>Jennifer Tufts and Tom Frank J. Acoust. Soc. Am. 101 , 3107
> >>
> >>
> >(1997) found
> >
> >
> >>that the accuracy of judging the number of talkers in a
> >>
> >>
> >multitalker mixture
> >
> >
> >>drops considerably when there are more than 3 talkers.  There
> >>
> >>
> >is also a
> >
> >
> >>report by David Huron (Music Perception, Vol. 19, No. 1
(2001)
> >>
> >>
> >pp. 1-64.,
> >
> >
> >>or on-line
> >>
> >>
> >>
>
>http://www.music-cog.ohio-state.edu/Huron/Publications/huron.voi
ce.leading.html
> >
> >
> >> ) that estimating the number of musical lines in
> >> polyphonic music worsens considerably after 3.  Some
anecdotal
> >>
> >>
> >evidence
> >
> >
> >>for this limit also comes from movie sound effect designers.
> >>
> >>
> >This is a
> >
> >
> >>citation from Walter Murch, a renown sound effect artist:
> >>
> >>
> >"There is a rule
> >
> >
> >>of thumb I use which is never to give the audience more than
> >>
> >>
> >two-and-a-half
> >
> >
> >>things to think about aurally at any one moment. Now, those
> >>
> >>
> >moments can
> >
> >
> >>shift very quickly, but if you take a five-second section of
> >>
> >>
> >sound and feed
> >
> >
> >>the audience more than two-and-a-half conceptual lines at the
> >>
> >>
> >same time,
> >
> >
> >>they can't really separate them out. There's just no way to
do
> >>
> >>
> >it, and
> >
> >
> >>everything becomes self-canceling." (cited from
> >>http://www.filmsound.org/murch/waltermurch.htm)
> >>
> >>Any thoughts, comments, and references relevant to this issue
> >>
> >>
> >are
> >
> >
> >>appreciated.
> >>
> >>
> >>
> >>-------------------------------------------------------------
> >>Valeriy Shafiro
> >>Communication Disorders and Sciences
> >>Rush University Medical Center
> >>Chicago, IL
> >>
> >>office (312) 942 - 3298
> >> lab    (312) 942 - 3316
> >>email: valeriy_shafiro@rush.edu
> >>
> >>
> >>
> >
> >
> >
>