Connectionists: Chomsky's apple

Brad Wyble bwyble at gmail.com
Thu Mar 9 17:01:42 EST 2023


I would argue that humans also have "patches of competence", or expertise,
and vast gulfs between them just like ChatGPT.  But what we have that
ChatGPT does not, is a means to approximate a metric of when we are
operating in an area of expertise (and for some those approximations are
looser than others..).

Even more importantly, we understand Q&A through a epistemic/moral
perspective which includes:
- that there is such a thing as being wrong
- that there is such a thing as value
- that there is value in being right
- Sometimes there is value in being wrong
- etc...

So I think that even once we develop some kind of causal reasoning engine
to allow a model to generalize somewhat accurately outside of its patches,
the harder question is to give the model an understanding of these deeper
aspects of interaction.  Otherwise it will always be easy to run them in
circles, break them out of their sandboxes, and prompt them to give stupid
answers.


On Thu, Mar 9, 2023 at 3:37 PM Dietterich, Thomas <tgd at oregonstate.edu>
wrote:

> ChatGPT’s errors reveal that its “understanding” of the world is not
> systematic but rather consists of patches of competence separated by
> regions of incompetence and incoherence. ChatCPT would be much stronger if
> it could fill in the gaps between those patches by appealing to general
> causal models. This raises two questions: (a) how could a system learn such
> causal models and (b) how could we test a system to determine whether it
> had succeeded.
>
>
>
> Researchers in the area of symbolic regression and causal discovery have
> an answer to (a): Learn a symbolic (e.g., DAG or differential equation)
> model. If we seek a fully connectionist account, how can we
> encourage/enforce systematicity in the knowledge acquired through deep
> learning? Regarding (b), does anyone have pointers to work on testing
> whether a connectionist system has acquired a systematic understanding of a
> causal relationship? This must go beyond testing (x,y) points, perhaps by
> verifying (Lipschitz) continuity?
>
>
>
> --Tom
>
>
>
> Thomas G. Dietterich, Distinguished Professor Voice: 541-737-5559
>
> School of Electrical Engineering              FAX: 541-737-1300
>
>   and Computer Science                        URL:
> eecs.oregonstate.edu/~tgd
>
> US Mail: 1148 Kelley Engineering Center
>
> Office: 2067 Kelley Engineering Center
>
> Oregon State Univ., Corvallis, OR 97331-5501
>
>
>
> *From:* Connectionists <connectionists-bounces at mailman.srv.cs.cmu.edu> *On
> Behalf Of *Stefan C Kremer
> *Sent:* Wednesday, March 8, 2023 17:44
> *To:* Terry Sejnowski <terry at salk.edu>
> *Cc:* Connectionists List <connectionists at cs.cmu.edu>
> *Subject:* Re: Connectionists: Chomsky's apple
>
>
>
> [This email originated from outside of OSU. Use caution with links and
> attachments.]
>
> One of the criticisms against John Searle’s argument (
> https://en.wikipedia.org/wiki/Chinese_room
> <https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fen.wikipedia.org%2Fwiki%2FChinese_room&data=05%7C01%7Ctgd%40oregonstate.edu%7C784658ebd66d4fa3f07308db2071adfb%7Cce6d05e13c5e4d6287a84c4a2713c113%7C0%7C0%7C638139444863338994%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=dah%2FU3oilsuzZ%2BzWZOtJVUwodjpPRjsjC3pkyPU9Ih4%3D&reserved=0>)
> has always been that it wouldn’t be possible to construct a book
> comprehensive enough to answer all the queries, or that it would take too
> long to produce an output.  Chat GPT shows that we have at least approached
> that limitation (perhaps not truly overcome it…yet).
>
>
>
> The question posed by Searle (and answered with a “yes” by Chomsky in his
> thinking about counterfactuals, causal explanation, and thinking) is: is
> there a difference between understanding and simulated understanding?
>
>
>
> I don’t know how we could ever answer this question (an to me that’s the
> important point), but it seems that Searle’s thought experiment becomes
> more relevant, now that a feasible implementation can be constructed, than
> when it was originally proposed.
>
>
>
> Stefan
>
>
>
>
>
> On Mar 8, 2023, at 12:25 PM, Terry Sejnowski <terry at salk.edu> wrote:
>
>
>
> CAUTION: This email originated from outside of the University of Guelph.
> Do not click links or open attachments unless you recognize the sender and
> know the content is safe. If in doubt, forward suspicious emails to
> IThelp at uoguelph.ca
>
>
> I have always been impressed with Chomsky's ability to use plausible
> arguments to make his case even when they were fallacious.
>
>
> https://www.nytimes.com/2023/03/08/opinion/noam-chomsky-chatgpt-ai.html#commentsContainer
>
> "Here’s an example. Suppose you are holding an apple in your hand. Now you
> let the apple go. You observe the result and say, “The apple falls.” That
> is a description. A prediction might have been the statement “The apple
> will fall if I open my hand.” Both are valuable, and both can be correct.
> But an explanation is something more: It includes not only descriptions and
> predictions but also counterfactual conjectures like “Any such object would
> fall,” plus the additional clause “because of the force of gravity” or
> “because of the curvature of space-time” or whatever. That is a causal
> explanation: “The apple would not have fallen but for the force of
> gravity.” That is thinking."
>
> His apple example sounded convincing, so I tested GPT on it:
>
> Me:  If I hold an apple in my hand and let it go what will happen?
>
> GPT:  The apple will fall to the ground due to the force of gravity.
>
> Me: What about any such object?
>
> GPT: Any object released from a person's hand will fall to the ground due
> to the force of gravity.
>
> Me:  What would happen if there wasn't a force of gravity?
>
> GPT:  If there were no force of gravity, objects would not be able to stay
> on the ground and would float away in all directions
>
> Chomsky defined thinking by the correct answers to these questions.
>
> Alas, the tide has turned.
>
> Terry
>
> -----
>
>
>


-- 
Brad Wyble
Professor of Psychology
Penn State University
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20230309/828d74d7/attachment.html>


More information about the Connectionists mailing list