<html><head><meta http-equiv="content-type" content="text/html; charset=utf-8"></head><body dir="auto"><div dir="ltr"></div><div dir="ltr">Stephen,</div><div dir="ltr"><br></div><div dir="ltr">On criteria (1)-(3), a high-end, mapping-equippped Roomba is far more plausible as a consciousness than GPT-3.</div><div dir="ltr"><br></div><div dir="ltr">1. The Roomba has a clearly defined wake-sleep cycle; GPT does not.</div><div dir="ltr">2. Roomba makes choices based on an explicit representation of its location relative to a mapped space. GPT lacks any consistent reflection of self; eg if you ask it, as I have, if you are you person, and then ask if it is a computer, it’s liable to say yes to both, showing no stable knowledge of self.</div><div dir="ltr">3. Roomba has explicit, declarative knowledge eg of walls and other boundaries, as well its own location. GPT has no systematically interrogable explicit representations.</div><div dir="ltr"><br></div><div dir="ltr">All this is said with tongue lodged partway in cheek, but I honestly don’t see what criterion would lead anyone to believe that GPT is a more plausible candidate for consciousness than any other AI program out there. </div><div dir="ltr"><br></div><div dir="ltr">ELIZA long ago showed that you could produce fluent speech that was mildly contextually relevant, and even convincing to the untutored; just because GPT is a better version of that trick doesn’t mean it’s any more conscious.</div><div dir="ltr"><br></div><div dir="ltr">Gary</div><div dir="ltr"><br><blockquote type="cite">On Feb 14, 2022, at 08:56, Stephen José Hanson <jose@rubic.rutgers.edu> wrote:<br><br></blockquote></div><blockquote type="cite"><div dir="ltr">
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
<p><font size="+1">this is a great list of behavior.. <br>
</font></p>
<p><font size="+1">Some biologically might be termed reflexive,
taxes, classically conditioned, implicit (memory/learning)...
all however would not be<br>
conscious in the several senses: (1) wakefulness-- sleep (2)
self aware (3) explicit/declarative.</font></p>
<p><font size="+1">I think the term is used very loosely, and I
believe what GPT3 and other AI are hoping to show signs of is
"self-awareness"..</font></p>
<p><font size="+1">In response to : "why are you doing that?",
"What are you doing now", "what will you be doing in 2030?"</font></p>
<p><font size="+1">Steve<br>
</font></p>
<p><font size="+1"><br>
</font></p>
<div class="moz-cite-prefix">On 2/14/22 10:46 AM, Iam Palatnik
wrote:<br>
</div>
<blockquote type="cite" cite="mid:CAFyaQg9J18EH5seELZcCymYgY8OXBh9oMO3V_=d-Nmc8f3gDDw@mail.gmail.com">
<meta http-equiv="content-type" content="text/html; charset=UTF-8">
<div dir="ltr">
<div>A somewhat related question, just out of curiosity.</div>
<div><br>
</div>
<div>Imagine the following:</div>
<div><br>
</div>
<div>- An automatic solar panel that tracks the position of the
sun.<br>
</div>
<div>
<div>- A group of single celled microbes with phototaxis that
follow the sunlight.</div>
<div>- A jellyfish (animal without a brain) that
follows/avoids the sunlight.</div>
<div>
- A cockroach (animal with a brain) that avoids the
sunlight.</div>
<div>- A drone with onboard AI that flies to regions of more
intense sunlight to recharge its batteries.<br>
</div>
<div>- A human that dislikes sunlight and actively avoids it.<br>
</div>
<div><br>
</div>
<div>Can any of these, beside the human, be said to be aware
or conscious of the sunlight, and why?</div>
<div>What is most relevant? Being a biological life form,
having a brain, being able to make decisions based on the
environment? Being taxonomically close to humans?<br>
</div>
<div><br>
</div>
<div><br>
</div>
<div><br>
</div>
<div><br>
</div>
<div><br>
</div>
<div><br>
</div>
</div>
</div>
<br>
<div class="gmail_quote">
<div dir="ltr" class="gmail_attr">On Mon, Feb 14, 2022 at 12:06
PM Gary Marcus <<a href="mailto:gary.marcus@nyu.edu" moz-do-not-send="true">gary.marcus@nyu.edu</a>> wrote:<br>
</div>
<blockquote class="gmail_quote" style="margin:0px 0px 0px
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">Also
true: Many AI researchers are very unclear about what
consciousness is and also very sure that ELIZA doesn’t have
it.<br>
<br>
Neither ELIZA nor GPT-3 have<br>
- anything remotely related to embodiment<br>
- any capacity to reflect upon themselves<br>
<br>
Hypothesis: neither keyword matching nor tensor manipulation,
even at scale, suffice in themselves to qualify for
consciousness.<br>
<br>
- Gary<br>
<br>
> On Feb 14, 2022, at 00:24, Geoffrey Hinton <<a href="mailto:geoffrey.hinton@gmail.com" target="_blank" moz-do-not-send="true">geoffrey.hinton@gmail.com</a>>
wrote:<br>
> <br>
> Many AI researchers are very unclear about what
consciousness is and also very sure that GPT-3 doesn’t have
it. It’s a strange combination.<br>
> <br>
> <br>
<br>
</blockquote>
</div>
</blockquote>
<div class="moz-signature">-- <br>
<img border="0" src="cid:part3.2ECE5112.986E94FB@rubic.rutgers.edu"></div>
</div></blockquote></body></html>