Re: Number of concurrent streams. (Al Bregman )


Subject: Re: Number of concurrent streams.
From:    Al Bregman  <bregman(at)HEBB.PSYCH.MCGILL.CA>
Date:    Mon, 3 May 2004 16:09:57 -0400

Dear Claire and list, I just made up the term "matching a standard" when I sent the last message. A discussion of this and other methods of measuring stream segregation is given in : Bregman, A. S., (1990/1994) "Auditory scene analysis.", in the section entitled, "How to measure streaming", pages 52-58. The method I described in my message is described in the sub-section "Pattern recognition" on p.56. Good luck. Al ----- Original Message ----- From: "Claire Piché" <clairepiche(at)VIF.COM> To: <AUDITORY(at)LISTS.MCGILL.CA> Sent: Monday, May 03, 2004 8:43 AM Subject: Re: Number of concurrent streams. > Dear M. Bregman > > Can you tell me more information about the method you called "matching a standard", Is it a syntagm of your own? Actually, i'm writing my thesis and a few days ago i started the part where your work is the main figure. Moreover, the second part of your letter reaches the heart of my work. > > > Claire Piché > Doctorat en Études et pratiques des arts > Université du Québec à Montréal > > > > > > > Al Bregman a écrit : > > >Dear Valeriy and List, > > > >(Sorry if you get this twice. Apparently the AUDITORY listserv > >didn't like my new email address and may or may not have sent > >this message to the list.) > > > >I don't think that the question about how many sound sources > >people can perceive in a mixture has a simple answer. > > > >First of all, it depends on what you mean by "perceive". If you mean how many sound sources they can report, then it depends on how much time they are given. The longer they listen, the more they can report. This could mean that they can pay close attention to only one sound at a time, but can shift their attention from one source to another. If you mean, "How many could they report if given a sample of each one in turn and asked whether they could detect it in the mixture, given all the time they needed?", the question becomes one about blending and segregation at a very basic level. (Let's call this last method, "matching a standard"). Often I haven't been able to hear a sound in a mixture until I knew what it was. > > > >Assuming that one is using the "matching a standard" method, > >then one's success will depend on what the sounds are and how intense they are. Obviously a weak sound may be hard to > >detect -- it may be masked (psychoacoustically or > >informationally) by the others. Even if this is not true, the similarity of the component sound sources plays an important role. Those researchers who have claimed that only a small number of sources (3 or 4) can be detected are all referring to sets of sounds that resemble one another, such as multiple talkers or singers, or rhythmically playing instruments. > > > >Consider the following set of sounds: > > - a person talking > > - randomly spaced hits on a bass drum (greatly attenuated) > > - an ambulance siren > > - jangling of a set of house keys, > > - a pure tone playing Morse code, > > - a person typing on an electric typewriter. > > - the sound of a motorcycle whizzing by (greatly attenuated) > > > >As long as the intensities were balanced appropriately, I think > >you would eventually detect all of them by the method of matching a standard. > >As many of the list members know, I believe that there is an > >early perceptual stage of assigning of links among the parts of the incoming mixture prior to further processing by the mechanisms that we call attention. If what we are asking about is whether this pre-attentive process has some limit concerning the number of discrete subsets (potential streams) it can form, we would have to observe its operation without any contribution from attention. I believe that > >this is impossible using the standard methods of psychoacoustics. Rather, it has to be addressed using a physiological approach. Some beginnings toward doing this have been carried out by Elyse Sussman and by Claude Alain (working independently), and there may be others who I don't know about. > > > >By the way, I referred to the output of the pre-attentive > >mechanism as "potential" streams because there is good reason to believe that top-down processes play a big role in determining the actually heard streams. > > > >Sorry the answer couldn't have been simpler. > > > >Al > > > >--------------------------------------------- > >Albert S. Bregman, > >Emeritus Professor > >Psychology Dept., McGill University > >1205 Docteur Penfield Avenue > >Montreal, Quebec > >Canada H3A 1B1 > > > >Office: > > Voice: +1 (514) 398-6103 > > Fax: +1 (514) 398-4896 > >---------------------------------------------. > > > > > > > >----- Original Message ----- > >From: "Valeriy Shafiro" <Valeriy_Shafiro(at)RUSH.EDU> > >To: <AUDITORY(at)LISTS.MCGILL.CA> > >Sent: Friday, April 30, 2004 3:26 PM > >Subject: Re: Computational ASA -- how many sources can humans > >perceive? > > > > > > > > > > > >>I would like to ask a further question: Do we, in fact, know > >> > >> > >how many > > > > > >>independent sound sources in a mixture humans can perceive? > >> > >> > >Thus far I > > > > > >>know of only one research report where human listeners were > >> > >> > >asked to > > > > > >>identify sound sources in a recorded "real-world" sound mixture > >> > >> > >(Ellis, D. > > > > > >>P. (1996). Prediction-driven computational auditory scene > >> > >> > >analysis). We > > > > > >>have been talking about this issue with Brian Gygi, and from > >> > >> > >the few > > > > > >>related reports that Brian found, it appears that humans may > >> > >> > >not be that > > > > > >>good in simultaneous perceiving independent sound sources. For > >> > >> > >instance, > > > > > >>Jennifer Tufts and Tom Frank J. Acoust. Soc. Am. 101 , 3107 > >> > >> > >(1997) found > > > > > >>that the accuracy of judging the number of talkers in a > >> > >> > >multitalker mixture > > > > > >>drops considerably when there are more than 3 talkers. There > >> > >> > >is also a > > > > > >>report by David Huron (Music Perception, Vol. 19, No. 1 (2001) > >> > >> > >pp. 1-64., > > > > > >>or on-line > >> > >> > >> > >http://www.music-cog.ohio-state.edu/Huron/Publications/huron.voi ce.leading.html > > > > > >> ) that estimating the number of musical lines in > >> polyphonic music worsens considerably after 3. Some anecdotal > >> > >> > >evidence > > > > > >>for this limit also comes from movie sound effect designers. > >> > >> > >This is a > > > > > >>citation from Walter Murch, a renown sound effect artist: > >> > >> > >"There is a rule > > > > > >>of thumb I use which is never to give the audience more than > >> > >> > >two-and-a-half > > > > > >>things to think about aurally at any one moment. Now, those > >> > >> > >moments can > > > > > >>shift very quickly, but if you take a five-second section of > >> > >> > >sound and feed > > > > > >>the audience more than two-and-a-half conceptual lines at the > >> > >> > >same time, > > > > > >>they can't really separate them out. There's just no way to do > >> > >> > >it, and > > > > > >>everything becomes self-canceling." (cited from > >>http://www.filmsound.org/murch/waltermurch.htm) > >> > >>Any thoughts, comments, and references relevant to this issue > >> > >> > >are > > > > > >>appreciated. > >> > >> > >> > >>------------------------------------------------------------- > >>Valeriy Shafiro > >>Communication Disorders and Sciences > >>Rush University Medical Center > >>Chicago, IL > >> > >>office (312) 942 - 3298 > >> lab (312) 942 - 3316 > >>email: valeriy_shafiro(at)rush.edu > >> > >> > >> > > > > > > >


This message came from the mail archive
http://www.auditory.org/postings/2004/
maintained by:
DAn Ellis <dpwe@ee.columbia.edu>
Electrical Engineering Dept., Columbia University