Connectionists: Yang and Piantodosi’s PNAS language system, semantics, and scene understanding

jose at rubic.rutgers.edu jose at rubic.rutgers.edu
Mon Jun 13 09:09:14 EDT 2022


I was thinking more like an RNN similar to work we had done in the 
2000s.. on syntax.

Stephen José Hanson, Michiro Negishi; On the Emergence of Rules in 
Neural Networks. Neural Comput 2002; 14 (9): 2245–2268. doi: 
https://doi.org/10.1162/089976602320264079

Abstract
A simple associationist neural network learns to factor abstract rules 
(i.e., grammars) from sequences of arbitrary input symbols by inventing 
abstract representations that accommodate unseen symbol sets as well as 
unseen but similar grammars. The neural network is shown to have the 
ability to transfer grammatical knowledge to both new symbol 
vocabularies and new grammars. Analysis of the state-space shows that 
the network learns generalized abstract structures of the input and is 
not simply memorizing the input strings. These representations are 
context sensitive, hierarchical, and based on the state variable of the 
finite-state machines that the neural network has learned. 
Generalization to new symbol sets or grammars arises from the spatial 
nature of the internal representations used by the network, allowing new 
symbol sets to be encoded close to symbol sets that have already been 
learned in the hidden unit space of the network. The results are counter 
to the arguments that learning algorithms based on weight adaptation 
after each exemplar presentation (such as the long term potentiation 
found in the mammalian nervous system) cannot in principle extract 
symbolic knowledge from positive examples as prescribed by prevailing 
human linguistic theory and evolutionary psychology.

On 6/13/22 8:55 AM, Gary Marcus wrote:
> – agree with Steve this is an interesting paper, and replicating it 
> with a neural net would be interesting; cc’ing Steve Piantosi.
> — why not use a Transformer, though?
> - it is however importantly missing semantics. (Steve P. tells me 
> there is some related work that is worth looking into). Y&P speaks to 
> an old tradition of formal language work by Gold and others that is 
> quite popular but IMHO misguided, because it focuses purely on syntax 
> rather than semantics.  Gold’s work definitely motivates learnability 
> but I have never taken it to seriously as a real model of language
> - doing what Y&P try to do with a rich artificial language that is 
> focused around syntax-semantic mappings could be very interesting
> - on a somewhat but not entirely analogous note, i think that  the 
> next step in vision is really scene understanding. We have techniques 
> for doing object labeling reasonably well, but still struggle wit 
> parts and wholes are important, and with relations more generally, 
> which is to say we need the semantics of scenes. is the chair on the 
> floor, or floating in the air? is it supporting the pillow? etc. is 
> the hand a part of the body? is the glove a part of the body? etc
>
> Best,
> Gary
>
>
>
>> On Jun 13, 2022, at 05:18, jose at rubic.rutgers.edu wrote:
>>
>> Again, I think a relevant project here would be to attempt to 
>> replicate with DL-rnn, Yang and Piatiadosi's PNAS language learning 
>> system--which is a completely symbolic-- and very general over the 
>> Chomsky-Miller grammer classes.   Let me know, happy to collaborate 
>> on something like this.
>>
>> Best
>>
>> Steve
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20220613/8575dda8/attachment.html>


More information about the Connectionists mailing list