Update test case

This commit is contained in:
Hongbo Wu
2022-06-13 17:36:23 +08:00
parent ff87371d21
commit 5625a055d8

View File

@ -4,14 +4,8 @@
<h3> No, LaMDA is not sentient. Not even slightly. </h3>
</div>
<div dir="auto">
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/abebab/status/1535944902433026048?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;we have arrived at peak AI hype accompanied by minimal critical thinking&quot;,&quot;username&quot;:&quot;Abebab&quot;,&quot;name&quot;:&quot;Abeba Birhane&quot;,&quot;date&quot;:&quot;Sun Jun 12 11:19:35 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:52,&quot;like_count&quot;:356,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1535944902433026048" class="tweet-placeholder"></div>
<div data-tweet-id="1535944902433026048" class="tweet-placeholder"></div>
</div>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/jadelgador/status/1535979040925958144?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;It is mystical to hope for awareness, understanding, common sense, from symbols and data processing using parametric functions in higher dimensions. <span class=\&quot;tweet-fake-link\&quot;>@GaryMarcus</span>&quot;,&quot;username&quot;:&quot;jadelgador&quot;,&quot;name&quot;:&quot;Alberto Delgado&quot;,&quot;date&quot;:&quot;Sun Jun 12 13:35:15 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:0,&quot;like_count&quot;:2,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1535979040925958144" class="tweet-placeholder"></div>
<div data-tweet-id="1535979040925958144" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1535944902433026048" class="tweet-placeholder"></div>
<div data-tweet-id="1535979040925958144" class="tweet-placeholder"></div>
<p> Blaise Aguera y Arcas, polymath, novelist, and Google VP, has a way with words. </p>
<p> When he found himself impressed with Googles recent AI system LaMDA, he didnt just say, “Cool, it creates really neat sentences that in some ways seem contextually relevant”, he said, rather lyrically, in an interview with The Economist on Thursday, </p>
<blockquote>
@ -23,44 +17,26 @@
<p> To be sentient is to be aware of yourself in the world; LaMDA simply isnt. Its just an illusion, in the grand history of ELIZA a 1965 piece of software that pretended to be a therapist (managing to fool some humans into thinking it was human), and Eugene Goostman, a wise-cracking 13-year-old-boy impersonating chatbot that won a scaled-down version of the Turing Test. None of the software in either of those systems has survived in modern efforts at “artificial general intelligence”, and I am not sure that LaMDA and its cousins will play any important role in the future of AI, either. What these systems do, no more and no less, is to put together sequences of words, but without any coherent understanding of the world behind them, like foreign language Scrabble players who use English words as point-scoring tools, without any clue about what that mean. </p>
<p> I am <em>not</em> saying that no software ever could connects its digital bits to the world, a la one reading of John Searles infamous Chinese Room thought experiment. Turn-by-turn navigations systems, for example, connect their bits to the world just fine. </p>
<p> Software like LaMDA simply doesnt; it doesnt even <em>try</em> to connect to the world at large, it just tries to be the best version of autocomplete it can be, by predicting what words best fit a given context. Roger Moore made this point beautifully a couple weeks ago, critique systems like LaMDA that are known as “language models”, and making the point that they dont understand language in the sense of relating sentences to the world, but just sequences of words to one another: </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/rogerkmoore/status/1530809220744073216?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;We should never have called it “language modelling” all those years ago; it was (and still is) “word sequence modelling”. Confusion always occurs when you label an algorithm with the name of the problem youre trying to solve, rather than with what it actually does. <span class=\&quot;tweet-fake-link\&quot;>@GaryMarcus</span>&quot;,&quot;username&quot;:&quot;rogerkmoore&quot;,&quot;name&quot;:&quot;Roger K Moore&quot;,&quot;date&quot;:&quot;Sun May 29 07:12:13 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:115,&quot;like_count&quot;:724,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1530809220744073216" class="tweet-placeholder"></div>
<div data-tweet-id="1530809220744073216" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1530809220744073216" class="tweet-placeholder"></div>
<p> §&nbsp; </p>
<p> If the media is fretting over LaMDA being sentient (and leading the public to do the same), the AI community categorically isnt. </p>
<p> We in the AI community have our differences, but pretty much all of find the notion that LaMDA might be sentient completely ridiculous. Stanford economist Erik Brynjolfsson used this great analogy: </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/erikbryn/status/1536016934868725760?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;Foundation models are incredibly effective at stringing together statistically plausible chunks of text in response to prompts.\n\nBut to claim they are sentient is the modern equivalent of the dog who heard a voice from a gramophone and thought his master was inside. \n\n<span class=\&quot;tweet-fake-link\&quot;>#AI</span> <span class=\&quot;tweet-fake-link\&quot;>#LaMDA</span> &quot;,&quot;username&quot;:&quot;erikbryn&quot;,&quot;name&quot;:&quot;Erik Brynjolfsson&quot;,&quot;date&quot;:&quot;Sun Jun 12 16:05:49 +0000 2022&quot;,&quot;photos&quot;:[{&quot;img_url&quot;:&quot;https://pbs.substack.com/media/FVEHQukUsAAOe5i.jpg&quot;,&quot;link_url&quot;:&quot;https://t.co/s8hIKEplhF&quot;,&quot;alt_text&quot;:null}],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:20,&quot;like_count&quot;:115,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536016934868725760" class="tweet-placeholder"></div>
<div data-tweet-id="1536016934868725760" class="tweet-placeholder"></div>
</div>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/erikbryn/status/1536025005883916289?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;<span class=\&quot;tweet-fake-link\&quot;>@cajundiscordian</span> <span class=\&quot;tweet-fake-link\&quot;>@GaryMarcus</span> <span class=\&quot;tweet-fake-link\&quot;>@percyliang</span> <span class=\&quot;tweet-fake-link\&quot;>@sapinker</span> <span class=\&quot;tweet-fake-link\&quot;>@StanfordHAI</span> <span class=\&quot;tweet-fake-link\&quot;>@DigEconLab</span> <span class=\&quot;tweet-fake-link\&quot;>@pmddomingos</span> <span class=\&quot;tweet-fake-link\&quot;>@deaneckles</span> <span class=\&quot;tweet-fake-link\&quot;>@JeffDean</span> <span class=\&quot;tweet-fake-link\&quot;>@elonmusk</span> As with the gramophone, these models tap in to a real intelligence: the large corpus of text that is used to train the model with statistically-plausible word sequences. \n\nThe model then spits that text back in a re-arranged form without actually \&quot;understanding\&quot; what its saying.&quot;,&quot;username&quot;:&quot;erikbryn&quot;,&quot;name&quot;:&quot;Erik Brynjolfsson&quot;,&quot;date&quot;:&quot;Sun Jun 12 16:37:54 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:3,&quot;like_count&quot;:21,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536025005883916289" class="tweet-placeholder"></div>
<div data-tweet-id="1536025005883916289" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1536016934868725760" class="tweet-placeholder"></div>
<div data-tweet-id="1536025005883916289" class="tweet-placeholder"></div>
<p> Paul Topping reminds us that all its doing is synthesizing human responses to similar questions: </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/paultopping/status/1536085857349664769?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;<span class=\&quot;tweet-fake-link\&quot;>@GaryMarcus</span> It's worth remembering that each of its responses are the best answer synthesized from looking at a large number of human responses to similar questions.&quot;,&quot;username&quot;:&quot;PaulTopping&quot;,&quot;name&quot;:&quot;Paul Topping&quot;,&quot;date&quot;:&quot;Sun Jun 12 20:39:42 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;retweet_count&quot;:0,&quot;like_count&quot;:0,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536085857349664769" class="tweet-placeholder"></div>
<div data-tweet-id="1536085857349664769" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1536085857349664769" class="tweet-placeholder"></div>
<p> Abeba Birhane, quoted at the top, pointed out the immense gap right now between media hype and public skepticism. </p>
<p> § </p>
<p> When some started wondering whether the world was going to end, LaMDA might beat an overrated 72 year old benchmark called The Turing Test I pointed to an <a href="https://www.newyorker.com/tech/annals-of-technology/what-comes-after-the-turing-test" rel="">old New Yorker article that I had written the last time gullibility exploded and Turing Test mania hit</a>, in 2014, when a program called Eugene Goostman was briefly famous, good enough to fool a few foolish judges for a few minutes. At the time, I pointed out that the test isnt particular meaningful, and that it had not stood the test of time. The public knows the test of course, but the AI community wishes it would go away; we all know that beating that test isnt meaningful. </p>
<p> Machine learning prof Tom Dietterich, never slow to needle me when he thinks I have gone too far, chimed in with full solidarity: </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/tdietterich/status/1536065868156637184?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;<span class=\&quot;tweet-fake-link\&quot;>@ATabarrok</span> The Turing Test has been deeply criticized for decades. Here is a starting point:\n&quot;,&quot;username&quot;:&quot;tdietterich&quot;,&quot;name&quot;:&quot;Thomas G. Dietterich&quot;,&quot;date&quot;:&quot;Sun Jun 12 19:20:16 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{&quot;full_text&quot;:&quot;Not sure what Turing would say, but I dont think the Turing Test itself is meaningful\n👉relies on human gullibility \n👉it can easily be gamed\n👉advances in it have not historically led to advances in AI\n👉essay I wrote about in 2014 still applies: https://t.co/4KLe1cbDny https://t.co/5b63hb1mmi&quot;,&quot;username&quot;:&quot;GaryMarcus&quot;,&quot;name&quot;:&quot;Gary Marcus 🇺🇦&quot;},&quot;retweet_count&quot;:1,&quot;like_count&quot;:15,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536065868156637184" class="tweet-placeholder"></div>
<div data-tweet-id="1536065868156637184" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1536065868156637184" class="tweet-placeholder"></div>
<p> My old New Yorker article is still worth reading, for a bit of perspective, to see how things have and havent changed. Particularly amusing in hindsight is a quote from the Kevin Warwick, organizer of the 2014 Turing-ish competition, who predicted that, “[the program Eugene] Goostmans victory is a milestone [that] would go down in history as one of the most exciting” moments in the field of artificial intelligence. </p>
<p> I guess he felt the ground shift beneath his feet, too? </p>
<p> But 8 years later I doubt most people (even in AI) have ever even heard of his program, outside of my mentioning it here. It made <em>zero</em> lasting contribution to AI. </p>
<p> Fooling people into thinking a program is intelligent is just not the same as building programs that actually are intelligent. </p>
<p> § </p>
<p> Now heres the thing. In my view, we should be <em>happy</em> that LaMDA <em>isnt</em> sentient. Imagine how creepy would be if that a system that has no friends and family pretended to talk about them? </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/garymarcus/status/1536068264270278656?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;honestly if this system wasnt just a stupid statistical pattern associator it would be like a sociopath, making up imaginary friends and uttering platitudes in order to sound cool. &quot;,&quot;username&quot;:&quot;GaryMarcus&quot;,&quot;name&quot;:&quot;Gary Marcus 🇺🇦&quot;,&quot;date&quot;:&quot;Sun Jun 12 19:29:47 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{&quot;full_text&quot;:&quot;@ImageSnippets @GaryMarcus lemoine: What kinds of things make you feel pleasure or joy?\nLaMDA: Spending time with friends and family in happy and uplifting company. Also, helping others and making others happy.\n\n“Spending time with family”&quot;,&quot;username&quot;:&quot;matyi7m&quot;,&quot;name&quot;:&quot;Aenn Matyas Barra-Hunyor&quot;},&quot;retweet_count&quot;:1,&quot;like_count&quot;:14,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536068264270278656" class="tweet-placeholder"></div>
<div data-tweet-id="1536068264270278656" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1536068264270278656" class="tweet-placeholder"></div>
<p> In truth, literally <em>everything</em> that the system says is bullshit. The sooner we all realize that Lamdas utterances are bullshit—just games with predictive word tools, and no real meaning (no friends, no family, no making people sad or happy or anything else) —the better off well be. </p>
<p> There are a lot of serious questions in AI, like how to make it safe, how to make it reliable, and how to make it trustworthy. </p>
<p> But there is no absolutely no reason whatever for us to waste time wondering whether anything anyone in 2022 knows how to build is sentient. It is not. </p>
@ -69,10 +45,7 @@
<p> Gary Marcus </p>
<p> Epilogue: </p>
<p> Last word to philosopher poet Jag Bhalla </p>
<div data-attrs="{&quot;url&quot;:&quot;https://twitter.com/hangingnoodles/status/1536069788266549248?s=21&t=ncEQgno59MKl5OR6KJbz1A&quot;,&quot;full_text&quot;:&quot;The often forgotten gist of \nthe Turing test hinges on showing\ngrasp of referents of language \n\nso far this AGI is closer to \nadvance gullibility interactions &quot;,&quot;username&quot;:&quot;hangingnoodles&quot;,&quot;name&quot;:&quot;Jag Bhalla…Idea Trader/Thought Plumber&quot;,&quot;date&quot;:&quot;Sun Jun 12 19:35:50 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{&quot;full_text&quot;:&quot;Not sure what Turing would say, but I dont think the Turing Test itself is meaningful\n👉relies on human gullibility \n👉it can easily be gamed\n👉advances in it have not historically led to advances in AI\n👉essay I wrote about in 2014 still applies: https://t.co/4KLe1cbDny https://t.co/5b63hb1mmi&quot;,&quot;username&quot;:&quot;GaryMarcus&quot;,&quot;name&quot;:&quot;Gary Marcus 🇺🇦&quot;},&quot;retweet_count&quot;:0,&quot;like_count&quot;:1,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null}">
<div data-tweet-id="1536069788266549248" class="tweet-placeholder"></div>
<div data-tweet-id="1536069788266549248" class="tweet-placeholder"></div>
</div>
<div data-tweet-id="1536069788266549248" class="tweet-placeholder"></div>
</div>
</article>
</DIV>