Connectionists: [EXT]Re: Fwd: Statistics versus “Understanding” in Generative AI.

Fernando Perez-Cruz fernando.perezcruz at sdsc.ethz.ch
Sun Feb 18 02:31:48 EST 2024


Gary,

Here is another short piece that I recently co-wrote to show the limitations of LLMs: https://www.bis.org/publ/bisbull83.htm.

It shows very clearly how they memorized and they cannot reason when given a math puzzle. 

Best,

Fernando 

---
Prof. Fernando Perez-Cruz
Titular Professor at the Computer Science Department (ETH Zürich)
https://scholar.google.com/citations?view_op=list_works&hl=en&hl=en&user=8FfrHw0AAAAJ




> On Feb 18, 2024, at 00:55, Gary Marcus <gary.marcus at nyu.edu> wrote:
> 
> GPT-4 is surely not reliably accurate; I gave many examples and pointed to Mitchell’s recent work as an example. The two empirical examinations Tom pointed to also note inaccuracies, as do many studies by Yejin Choi, Subbarao Kambhampati, and so on.
> 
> OpenAI regularly updates their systems with undisclosed mechanisms and addresses some publicly discussed examples. (it is also stochastic, so even on questions it gets right it sometimes makes errors).
> 
> But their patches are rarely if ever robust; it is absurd to suggest that GPT-4 is accurate, full stop, with no qualification. 
> 
>> On Feb 17, 2024, at 3:47 PM, Fellous, Jean-Marc - (fellous) <fellous at arizona.edu> wrote:
>> 
>> 
>> Hi Gary,
>>  
>> Yes, I meant the GPT-4 output. If I understand your answer, it is accurate.
>> Then my point is that whether GPT ‘understands’ or fakes an understanding may be secondary to the fact that it made us understand (to some good? extent) your point of view on AI. It may not ‘have’ an understanding, but it generates one.
>> That is amazing.
>> What would be good now, is if there was a (relatively safe) feedback mechanism that would allow for a (human) reinforcement of whatever lead to that output. Moving from a purely feedforward GPT (query->answer) to a recurrent one (answer + human->GPT inner workings)…
>>  
>> Best,
>> Jean-Marc
>>  
>> From: Gary Marcus <gary.marcus at nyu.edu> 
>> Sent: Friday, February 16, 2024 9:16 AM
>> To: Fellous, Jean-Marc - (fellous) <fellous at arizona.edu>
>> Cc: Terry Sejnowski <terry at snl.salk.edu>; connectionists at cs.cmu.edu
>> Subject: Re: [EXT]Re: Connectionists: Fwd: Statistics versus “Understanding” in Generative AI.
>>  
>> External Email
>> 
>> On your first question: Is what accurate? There is no substantive claim to be tested in Sejnowski’s schoolyard taunt. 
>>  
>> As for my position on GPT-4, one place to start: All my predictions of GPT-4 were poublished before its public release (and before I had access); I think all were correct; I don’t think I revised any. You can read them here: https://garymarcus.substack.com/p/what-to-expect-when-youre-expecting <https://urldefense.proofpoint.com/v2/url?u=https-3A__garymarcus.substack.com_p_what-2Dto-2Dexpect-2Dwhen-2Dyoure-2Dexpecting&d=DwMGaQ&c=slrrB7dE8n7gBJbeO0g-IQ&r=wQR1NePCSj6dOGDD0r6B5Kn1fcNaTMg7tARe7TdEDqQ&m=Tk9Tsq2lRHVKW_v2w_dSieqtGjS5HRGpyvPdkcTb8w_iY7lnlguK4xlwUl1ueOA_&s=D-WgPlGV1AzE_9ICCZIaS-C9yGeU0zTkuBChEperbZ8&e=>
>> 
>> 
>> On Feb 16, 2024, at 08:10, Fellous, Jean-Marc - (fellous) <fellous at arizona.edu <mailto:fellous at arizona.edu>> wrote:
>> 
>> 
>> Is it accurate though?
>> Some of us sure could us a concise summary of everyone's position/opinion!
>> Thanks,
>> Jean-Marc
>>  
>>  
>>  
>> From: Connectionists <connectionists-bounces at mailman.srv.cs.cmu.edu <mailto:connectionists-bounces at mailman.srv.cs.cmu.edu>> on behalf of Gary Marcus <gary.marcus at nyu.edu <mailto:gary.marcus at nyu.edu>>
>> Sent: Friday, February 16, 2024 5:11 AM
>> To: Terry Sejnowski <terry at snl.salk.edu <mailto:terry at snl.salk.edu>>
>> Cc: connectionists at cs.cmu.edu <mailto:connectionists at cs.cmu.edu> <connectionists at cs.cmu.edu <mailto:connectionists at cs.cmu.edu>>
>> Subject: [EXT]Re: Connectionists: Fwd: Statistics versus “Understanding” in Generative AI.
>>  
>> External Email
>> 
>> Always with the childish ad hominem. 
>>  
>> This field is something else.
>> 
>> 
>> On Feb 15, 2024, at 11:09 PM, Terry Sejnowski <terry at snl.salk.edu <mailto:terry at snl.salk.edu>> wrote:
>> 
>>  GPT-4 is a moving target that may eventually replace Gary.
>> 
>> Terry
>> 
>> -------
>> 
>> <3ttu1d7S0cx4Qvpk.png>
>>  
>> <uGPqolaZPAByhZbS.png>
>> 
>> The request you've made is similar to one I've already fulfilled. The image generated earlier depicts an empty room without any elephant present, adhering to your specific instructions. If you need any further adjustments or have another request, feel free to let me know!
>> ------
>> 
>> 
>> 
>> Write a critique of ChatGPT in the style of Gary Marcus
>> 
>> ChatGPT
>> Writing a critique of ChatGPT in the style of Gary Marcus, a scientist known for his clear-eyed scrutiny of artificial intelligence's capabilities and limitations, presents an intriguing challenge. Marcus, with his deep understanding of both the potential and pitfalls of AI, often emphasizes the importance of grounding AI advancements in robust, generalized learning principles, rather than overreliance on data and statistical patterns.
>> 
>> Critique of ChatGPT in the Style of Gary Marcus:
>> 
>> ChatGPT, while a marvel of contemporary AI technology, epitomizes the quintessential limitations that besiege deep learning models. At its core, ChatGPT is a linguistic colossus, built upon the edifice of the transformer architecture, yet it is fundamentally constrained by the inherent limitations of its training methodology and underlying principles. The brilliance of generating coherent, contextually appropriate responses is overshadowed by its occasional lapses into confabulation, a glaring testament to its lack of true understanding.
>> 
>> Gary Marcus would likely point out that ChatGPT, for all its linguistic dexterity, lacks the nuanced grasp of the world that humans possess. It can mimic the form of human knowledge and discourse, but it doesn't truly comprehend the content in the way a human does. This distinction is crucial; it is the boundary between a sophisticated parrot and an entity capable of genuine thought and understanding. ChatGPT's responses, no matter how convincing, are ultimately derived from patterns in data it has been fed, not from an intrinsic understanding of those patterns.
>> 
>> Furthermore, Marcus might criticize ChatGPT's reliance on sheer volume of data as a substitute for a more nuanced, rule-based understanding of language and the world. This data-driven approach can lead to impressive performances in many scenarios but falls short when faced with novel situations or when required to apply common sense reasoning that hasn't been explicitly encoded in its training data. The model's occasional incoherence or its propensity to generate plausible but factually incorrect responses are symptomatic of this deeper issue.
>> 
>> Another point of critique would likely focus on the opacity of ChatGPT's decision-making process. Despite recent advances in making AI models more interpretable, understanding how ChatGPT arrives at a particular response remains a challenge. This opacity complicates efforts to debug or improve the model, especially when it produces erroneous or nonsensical outputs. For AI to truly advance, Marcus might argue, it must not only perform well but do so in a manner that is transparent and understandable to human overseers.
>> 
>> In the spirit of Marcus's advocacy for a hybrid approach to AI development, one could argue that integrating ChatGPT's statistical prowess with more structured, rule-based systems could ameliorate some of its shortcomings. By grounding the model's responses in a framework of logical reasoning and factual correctness, one might achieve a more reliable, genuinely intelligent system that better approximates human cognition.
>> 
>> In conclusion, while ChatGPT represents a significant achievement in natural language processing, its limitations underscore the need for a more comprehensive approach to AI. An approach that incorporates the strengths of both deep learning and symbolic AI, coupled with a commitment to transparency and understandability, may pave the way for the development of AI systems that are not only more capable but also more aligned with the nuanced complexities of human intelligence and reasoning.
>> 
>>  
>> 
>> -------- Forwarded Message --------
>> Subject:
>> Connectionists: Statistics versus “Understanding” in Generative AI.
>> Date:
>> Tue, 13 Feb 2024 09:54:48 -0800
>> From:
>> Gary Marcus <gary.marcus at nyu.edu> <mailto:gary.marcus at nyu.edu>
>> To:
>> Weng, Juyang <weng at msu.edu> <mailto:weng at msu.edu>
>> CC:
>> connectionists at mailman.srv.cs.cmu.edu <mailto:connectionists at mailman.srv.cs.cmu.edu>
>> 
>> ...
>> 
>> A few sample visuals from those below (read the essays for sources). The argument in brief is that the systems do fine in statistically canonical situations, but lack depth in their comprehension of complex syntax, negation, unusual circumstances, etc. Much more context (and many more examples) in the essays themselves.
>>  
>> Gary
>> 
>> <a0ET6JyMvGOjno0L.png>
>>  
>> 
>> Gary
>>  
>> <https__substack-post-media.s3.amazonaws.com_public%2Fimages_ab5345a6-2f04-463d-a53b-789115a1fdb7_851x1024.jpeg>
>> <IMG_5914.jpeg>
>> <IMG_3818.jpeg>
>> <IMG_3811.jpeg>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20240218/be554768/attachment.html>


More information about the Connectionists mailing list