
VANCOUVER, CANADA -
In hindsight, ChatGPT may come to be seen as the greatest publicity stunt in AI history, an intoxicating glimpse at a future that may actually take years to realize - kind of like a 2012-vintage driverless car demo, but this time with a foretaste of an ethical guardrail that will take years to perfect.
What ChatGPT delivered in spades that its predecessors like Microsoft Tay - released March 23, 2016 and withdrawn March 24 for toxic behavior - and Meta’s Galactica - released November 16, 2022 and withdrawn November 18 - could not was an illusion, a sense that the problem of toxic spew was finally coming under control. ChatGPT rarely says anything overtly racist, and simple requests for antisemitism and outright lies are often rebuffed.
Indeed, at times it can seem so politically correct that the right wing has become enraged. Elon Musk, for one, has expressed concern that the system has become an agent of wokeness:
However - as is often the case - the reality is actually more nuanced and complex than that.
The thing to remember - as I have emphasized many times - is that ChatGPT has no idea of what it’s talking about. Mistakenly believing that ChatGPT has any moral views at all is pure, unadulterated anthropomorphism.
From a technical standpoint, the thing that allegedly made ChatGPT so much better than Galactica - which was released a couple weeks earlier only to be withdrawn three days later
The content herein is subject to copyright by The Yuan. All rights reserved. The content of the services is owned or licensed to The Yuan. The copying or storing of any content for anything other than personal use is expressly prohibited without prior written permission from The Yuan, or the copyright holder identified in the copyright notice contained in the content.




