Connectionists: Yang and Piantodosi’s PNAS language system, semantics, and scene understanding

Stephen Jose Hanson stephen.jose.hanson at rutgers.edu
Mon Jun 13 11:45:16 EDT 2022


It would have to be updated with with DL-RNN or LSTMs..

S

On 6/13/22 9:13 AM, Gary Marcus wrote:
 I do remember the work :) Just generally Transformers seem more effective; a careful comparison between Y&P, Transformers, and your RNN approach, looking at generalization to novel words, would indeed be interesting.
Cheers,
Gary

On Jun 13, 2022, at 06:09, jose at rubic.rutgers.edu<mailto:jose at rubic.rutgers.edu> wrote:



I was thinking more like an RNN similar to work we had done in the 2000s.. on syntax.

Stephen José Hanson, Michiro Negishi; On the Emergence of Rules in Neural Networks. Neural Comput 2002; 14 (9): 2245–2268. doi: https://doi.org/10.1162/089976602320264079<https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Furldefense.com%2Fv3%2F__https%3A%2F%2Fdoi.org%2F10.1162%2F089976602320264079__%3B!!BhJSzQqDqA!WCsRlT1zpBKD3ai8Ov_I79iH_HCdTlAMymGIe2ZsIdTnfZawzlMQNGZWisMjmcLBgH6SbBUZ6rtr_exEspS4Igo%24&data=05%7C01%7Cstephen.jose.hanson%40rutgers.edu%7Cdf1d37eb5a494e8cca8108da4d4261ba%7Cb92d2b234d35447093ff69aca6632ffe%7C1%7C0%7C637907244767470752%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=DiJWII1n%2F24jAzC%2B47UpHF1%2FdhfCHKRia%2BTdl2jZsd0%3D&reserved=0>

Abstract
A simple associationist neural network learns to factor abstract rules (i.e., grammars) from sequences of arbitrary input symbols by inventing abstract representations that accommodate unseen symbol sets as well as unseen but similar grammars. The neural network is shown to have the ability to transfer grammatical knowledge to both new symbol vocabularies and new grammars. Analysis of the state-space shows that the network learns generalized abstract structures of the input and is not simply memorizing the input strings. These representations are context sensitive, hierarchical, and based on the state variable of the finite-state machines that the neural network has learned. Generalization to new symbol sets or grammars arises from the spatial nature of the internal representations used by the network, allowing new symbol sets to be encoded close to symbol sets that have already been learned in the hidden unit space of the network. The results are counter to the arguments that learning algorithms based on weight adaptation after each exemplar presentation (such as the long term potentiation found in the mammalian nervous system) cannot in principle extract symbolic knowledge from positive examples as prescribed by prevailing human linguistic theory and evolutionary psychology.


On 6/13/22 8:55 AM, Gary Marcus wrote:
– agree with Steve this is an interesting paper, and replicating it with a neural net would be interesting; cc’ing Steve Piantosi.
— why not use a Transformer, though?
- it is however importantly missing semantics. (Steve P. tells me there is some related work that is worth looking into). Y&P speaks to an old tradition of formal language work by Gold and others that is quite popular but IMHO misguided, because it focuses purely on syntax rather than semantics.  Gold’s work definitely motivates learnability but I have never taken it to seriously as a real model of language
- doing what Y&P try to do with a rich artificial language that is focused around syntax-semantic mappings could be very interesting
- on a somewhat but not entirely analogous note, i think that  the next step in vision is really scene understanding. We have techniques for doing object labeling reasonably well, but still struggle wit parts and wholes are important, and with relations more generally, which is to say we need the semantics of scenes. is the chair on the floor, or floating in the air? is it supporting the pillow? etc. is the hand a part of the body? is the glove a part of the body? etc

Best,
Gary



On Jun 13, 2022, at 05:18, jose at rubic.rutgers.edu<mailto:jose at rubic.rutgers.edu> wrote:


Again, I think a relevant project here  would be to attempt to replicate with DL-rnn, Yang and Piatiadosi's PNAS language learning system--which is a completely symbolic-- and very general over the Chomsky-Miller grammer classes.   Let me know, happy to collaborate on something like this.

Best

Steve

--
[cid:part3.207D6B48.20BF6C56 at rutgers.edu]
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20220613/eb3cf1e4/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.png
Type: image/png
Size: 34455 bytes
Desc: signature.png
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20220613/eb3cf1e4/attachment.png>


More information about the Connectionists mailing list