[ACT-R-users] Models of listening to continuous speech?

Bonnie John bej at cs.cmu.edu
Wed Aug 12 11:52:13 EDT 2009


This helps a LOT.
Thanks,
Bonnie


Martin Greaves wrote:
> Hi Bonnie
>
> Further to Richard's comments, one of the key issues we were interested 
> in was rehearsal of auditory items.  We attempted to model the word 
> length effect in ACT-R using the audicon to represent a phonological 
> loop-like structure.  Whereas the Huss & Byrne model used declarative 
> memory for item storage and decay, and encoded list position within the 
> item chunk, we wanted to see whether a phonological loop could be 
> achieved using the intrinsic mechanisms of the audicon to account for 
> order and decay.  However, we found that maintaining any form of 
> rehearsal in the audicon to be difficult to implement.
>
> One of the key problems is determining when the end of a list of items 
> has been achieved, since ACT-R places 'rehearsed' items back in the 
> audicon as unattended items.  Hence, once an item has been rehearsed, it 
> is added to the end of the existing list of items with nothing to 
> discriminate it from items in the current (unrehearsed) list of items.  
> Consequently, the audicon does not allow discrimination of lists.  In 
> order to get around this problem and determine when the end of the list 
> had been reached we added an 'end' item.  As Richard said, this was 
> argued as being analogous to a drawing of breath or a pause prior to any 
> further rehearsal of items.
>
> A further problem with the model was that it failed to reproduce serial 
> order effects.  Although we obtained a good fit for word length, 
> particularly for one-syllable words, we found that serial position 
> curves were 'm' shaped rather than displaying extended primacy and 
> single-item recency.  This was because items simply dropped out of 
> memory if they were not rehearsed within the decay period, with no 
> possibility of recovery due to the all-or-nothing decay in the audicon.  
> Without intending to be in any way critical, this was not really 
> surprising given the operation of the audicon.  However, rehearsal would 
> not appear to be an issue in your case, so recall from the audicon 
> should simply represent an extended recency effect.
>
> In the running memory span models we took a slightly different approach 
> to the model.  We assumed that storage might utilise a range of 
> available short-term memory structures, with recall from whichever 
> structures still contained items at recall.  This allowed continued use 
> of the audicon but brings declarative  memory into play, since a 
> spin-off from sub-vocal rehearsal is an item chunk in DM.  The audicon 
> was considered it to operate simply as an extended buffer containing the 
> last 2-3 items presented.  Thus, we recalled whatever was available from 
> the audicon at the point of recall, with order implicit in recall, prior 
> to recall of earlier items from declarative memory.  Using this approach 
> we were able to account for a number of possible rehearsal strategies 
> and what we termed rehearsal 'micro-strategies' (cf. Gray & Boehm-Davis, 
> 2000) reproducing a range of patterns of recall observed empirically.  
> Were were also able to account for patterns of delayed recall when items 
> in the audicon had long since decayed through recall from declarative 
> memory.
>
> Hope this is of help,
>
> Martin
>
>
>
>
>
>
> Richard M Young wrote:
>   
>> Hello Bonnie,
>>
>> Some years ago, following on from David Huss's work with Mike Byrne, 
>> Martin Greaves and I played around with simple models using the 
>> audicon to deal with lists of isolated words in a short-term memory 
>> experiment.  At least in our hands, the audicon was assumed to hold 
>> "words" in some unspecified phonemic/phonological/articulatory code, 
>> which therefore required an access to DM/LTM in order to retrieve the 
>> word as a lexical item.
>>
>> As I remember it -- which is not well -- we found it a bit clunky to 
>> use the audicon to deal with ephemeral material spread out in time. 
>> With no disrespect to those who had worked on it, I think it's fair to 
>> say that not much thought had been given to the underlying design of a 
>> transducer dealing with speech sounds.  For example, I *think* we had 
>> to add an END signal to indicate the silence following the last item 
>> in a list.
>>
>> Martin later build partial models of the running memory span task 
>> using a similar approach, and included them in his PhD thesis.  He may 
>> be able to add to what I'm saying here.
>>
>> Word recognition in speech is a well-studied area in cognitive 
>> psychology, and there are some good (ad hoc) models around, and I 
>> believe a reasonable amount of consensus.  It had crossed my mind from 
>> time to time that it would be an interesting area to bring into 
>> contact with Act-R.  There are of course Act-R models of the lexical 
>> retrieval stage itself.
>>
>> Good luck!
>>
>> ~ Richard
>>
>> At 11:18 -0400 11/8/09, Bonnie John wrote:
>>     
>>> Folks,
>>>
>>> Anyone have an models using ACT-R's Audition Module to listen to and
>>> comprehend continuous speech?
>>> If not continuous speech, how about short phrases or anything other than
>>> tones?
>>> Any experience using the Audition Module at all?
>>>
>>> I'm asking because I'd like to make models of ACT-R using the JAWS
>>> screen reader to navigate a web site and compare it to ACT-R models of
>>> visually navigating a web site. I'd like to read papers or speak with
>>> someone who has had experience with the Audition Module to get a
>>> heads-up on peaks and pitfalls.
>>>
>>> I did browse the Publications page, but could only find papers that use
>>> vision (but I could have missed the listening ones - apologies if I 
>>> did).
>>>
>>> Thanks,
>>> Bonnie
>>>       
> _______________________________________________
> ACT-R-users mailing list
> ACT-R-users at act-r.psy.cmu.edu
> http://act-r.psy.cmu.edu/mailman/listinfo/act-r-users
>
>   




More information about the ACT-R-users mailing list