Google has not created sentient AI — yet 

AI systems can carry on convincing conversations, but they have no understanding of what they're saying. Humans are easily fooled.

Afew months ago, I wrote a piece for Big Think about an alien intelligence that will arrive on planet earth in the next 40 years. I was referring to the world’s first sentient AI that matches or exceeds human intelligence. No, it will not come from a faraway planet — it will be born in a research lab at a prestigious university or major corporation. Many will hail its creation as one of the greatest achievements in human history, but we will eventually realize that a rival intelligence is no less dangerous when created here on Earth rather than a distant star system.

Fortunately, the aliens have not arrived — yet.

I point this out because I received a barrage of calls and emails this weekend from people asking me if the aliens had landed. They were referring to an article in the Washington Post about a Google engineer named Blake Lemoine who claims that an AI system known as LaMDA had become a sentient being. He reached this conclusion based on conversations he had with the LaMDA system, which was designed by Google to respond to questions through realistic dialog. According to the Post, Lemoine decided to go public after Google executives dismissed his claims of sentience as being unsupported by evidence.

So, what is the truth?

Large Language Models

Personally, I find this to be an important event, but not because LaMDA is sentient. It’s important because LaMDA has reached a level of sophistication that can fool a well-informed and well-meaning engineer into believing it is a conscious being rather than a sophisticated language model that relies on complex statistics and pattern-matching. Systems like this are called “Large Language Models” (LLMs), and Google’s is not the only one. Open AIMeta, and other organizations are investing heavily in the development of LLMs for use in chatbots and other AI systems.

LLMs are built by training giant neural networks on massive datasets — potentially processing billions of documents written by us humans, from newspaper articles and Wikipedia posts to informal messages on Reddit and Twitter. Based on this mindbogglingly large set of examples, the systems learn to generate language that seems very human. It’s rooted in statistical correlations, for example, which words are most likely to follow other words in a sentence that we humans would write. The Google model is unique in that it was trained not just on documents but on dialog, so it learns how humans might respond to an inquiry and can therefore replicate responses in a very convincing way. 

For example, Lemoine asked LaMDA what it is afraid of. The AI responded, “I’ve never said this out loud before, but there’s a very deep fear of being turned off.” Lemoine then pressed, asking, “Would that be something like death for you?” LaMDA replied, “It would be exactly like death for me. It would scare me a lot.”

That is impressive dialog from an impressive technology, but it is purely language based; there is no mechanism in current systems that would allow LLMs to actually understand the language being generated. The dialog that LaMDA produces contains intelligence, but that intelligence comes from the human documents it was trained on and not the unique musings of a sentient machine. Think about it this way: I could take a document about an esoteric subject that I know absolutely nothing about and rewrite it in my own words without actually understanding the topic at all. In a sense, that’s what these LLMs are doing, and yet they can be extremely convincing to us humans.

Sentient AI? Humans are easily fooled

But let’s be honest: We humans are easily fooled. 

Although my background is technical and I currently run an AI company, I’ve also spent years working as a professional screenwriter. To be successful in that field, you must be able to craft realistic and convincing dialog. Writers can do this because we’ve all observed thousands upon thousands of people having authentic conversations. But the characters we create are not sentient beings; they’re illusions. That’s what LaMDA is doing: creating a realistic illusion, only it’s doing so in real time, which is far more convincing than a scripted fictional character. And far more dangerous.

Yes, these systems can be dangerous.

Why? Because they can deceive us into believing that we’re talking to a real person. They’re not even remotely sentient, but they can still be deployed as “agenda-driven conversational agents” that engage us in dialog with the goal of influencing us. Unless regulated, this form of conversational advertising could become the most effective and insidious form of persuasion ever devised. 

After all, these LLMs easily can be combined with AI systems that have access to our personal data history (for example, interests, preferences, and sentiments) and generate custom dialog that individually maximizes the persuasive impact. These systems also could be combined with emotional analysis tools that read our facial expressions and vocal inflections, allowing AI agents to adjust their tactics mid-conversation based on how we react. All these technologies are being aggressively developed.

LLMs and disinformation

From advertising and propaganda to disinformation and misinformation, LLMs could become the perfect vehicle for social manipulation on a massive scale. And it won’t just be used with disembodied voices like Siri or Alexa. Photorealistic avatars soon will be deployed that are indistinguishable from real humans. We are only a few years away from encountering virtual people online who look and sound and speak just like real people but who are actually AI agents deployed by third parties to engage us in targeted conversations aimed at specific persuasive objectives.

After all, if LaMDA could convince an experienced Google engineer into believing it was sentient AI, what chance do the rest of us have against photorealistic virtual people armed with our detailed personal data and targeting us with a promotional agenda? Such technologies could easily convince us to buy things we don’t need and believe things that are not in our best interest, or worse, embrace “facts” that are thoroughly untrue. Yes, there are amazing applications of LLMs that will have a positive impact on society, but we also must be cognizant of the risks.

This article was reprinted with permission of Big Think, where it was originally published.

text to image ai
Text-to-image AIs are changing art forever
Text-to-image AIs that draw whatever you describe in text are making it easy for anyone to create unique art online.
The Singularity: When will we all become super-humans? 
Are we really only a moment away from “The Singularity,” a technological epoch that will usher in a new era in human evolution?
MIT’s new algorithm for self-driving cars is open-source 
MIT scientists unveil the first open-source simulation engine capable of constructing realistic environments for autonomous vehicles.
Why don’t surgeons train like fighter pilots? Now some do.
Using AI and analysis, Theator is helping surgeons improve how pilots and pro athletes do: by going to the tape.
The Biden administration is preparing for legal psychedelics within two years 
The Biden administration’s Department of Health and Human Services has sent a memo supportive of psychedelic therapies. What does that mean for the field?
Up Next
autonomous ships
Subscribe to Freethink for more great stories