EletiofeDon’t Ask Dumb Robots If AI Will Destroy Humanity

Don’t Ask Dumb Robots If AI Will Destroy Humanity

-

- Advertisment -

Earlier this month, several prominent outlets carried news that artificial intelligence will not pose a danger to humanity. The source of this reassuring news? A bunch of humanoid robot heads connected to simple chatbots.

The news stories sprang from a panel at a United Nations conference in Geneva called AI for Good, where several humanoids appeared alongside their creators. Reporters were invited to ask questions to the robots, which included Sophia, a machine made by Hanson Robotics that has gained notoriety for appearing on talk shows and even, bizarrely, gaining legal status as a person in Saudi Arabia.

The questions included whether AI would destroy humanity or steal jobs. Their replies were made possible by chatbot technology, somewhat similar to that which powers ChatGPT. But despite the well-known limitations of such bots, the robots’ replies were reported as if they were the meaningful opinions of autonomous, intelligent entities.

Why did this happen? Robots that can visually mimic human expressions trigger an emotional response in onlookers because we are so primed to pick up on such cues. But allowing what is nothing more than advanced puppetry to disguise the limitations of current AI can confuse people trying to make sense of the technology or of recent concerns about problems it may cause. I was invited to the Geneva conference, and when I saw Sophia and other robots listed as “speakers,” I lost interest.

It’s frustrating to see such nonsense at a time when more trustworthy experts are warning about current and future risks posed by AI. Machine learning algorithms are already exacerbating social biases, spewing disinformation, and increasing the power of some of the world’s biggest corporations and governments. Leading AI experts worry that the pace of progress may produce algorithms that are difficult to control in a matter of years.

Hanson Robotics, the company that makes Sophia and other lifelike robots, is impressively adept at building machines that mimic human expressions. Several years ago, I visited the company’s headquarters in Hong Kong and met with founder David Hanson, who previously worked at Disney, over breakfast. The company’s lab was like something from Westworld or Blade Runner, with unplugged robots gazing sadly into the middle distance, shriveled faces flopped on shelves, and prototypes stuttering the same words over and over in an infinite loop.

Photograph: Will Knight

Hanson and I talked about the idea of adding real intelligence to these evocative machines. Ben Goertzel, a well-known AI researcher and the CEO of SingularityNET, leads an effort to apply advances in machine learning to the software inside Hanson’s robots that allows them to respond to human speech.

The AI behind Sophia can sometimes provide passable responses, but the technology isn’t nearly as advanced as a system like GPT-4, which powers the most advanced version of ChatGPT and cost more than $100 million to create. And of course even ChatGPT and other cutting-edge AI programs cannot sensibly answer questions about the future of AI. It may be best to think of them as preternaturally knowledgeable and gifted mimics that, although capable of surprisingly sophisticated reasoning, are deeply flawed and have only a limited “knowledge” of the world.

Sophia and company’s misleading “interviews” in Geneva are a reminder of how anthropomorphizing AI systems can lead us astray. The history of AI is littered with examples of humans overextrapolating from new advances in the field.

In 1958, at the dawn of artificial intelligence, The New York Times wrote about one of the first machine learning systems, a crude artificial neural network developed for the US Navy by Frank Rosenblatt, a Cornell psychologist. “The Navy revealed the embryo of an electronic computer today that it expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence,” the Times reported—a bold statement about a circuit capable of learning to spot patterns in 400 pixels.

If you look back at the coverage of IBM’s chess-playing Deep Blue, DeepMind’s champion Go player AlphaGo, and many of the past decade’s leaps in deep learning—which are directly descended from Rosenblatt’s machine—you’ll see plenty of the same: people taking each advance as if it were a sign of some deeper, more humanlike intelligence.

That’s not to say that these projects—or even the creation of Sophia—were not remarkable feats, or potentially steps toward more intelligent machines. But being clear-eyed about the capabilities of AI systems is important when it comes to gauging progress of this powerful technology. To make sense of AI advances, the least we can do is stop asking animatronic puppets silly questions.

Latest news

Top Home Chef Promo Codes for May 2026

Out of the dozens of services I’ve tested, Home Chef is my favorite meal kit service for beginner cooks....

Sealy Promo Codes: $100 Off

Sealy is a mattress brand that is tried and true for many people, given that it has been around...

Newegg Promo Code: 10% Off in May 2026

Listen up, nerds. Newegg currently has promo codes and deals on gently used, refurbished, new and hard-to-find electronics, gaming...

eBay Coupons: 20% Off in May 2026

Long before we had Amazon or Facebook marketplace, or thousands of other online retailers, we had eBay. And now,...
- Advertisement -

Petlibro Offers: 60% Off in May

As the pet tech writer here on the WIRED Reviews team, I’ve tested over 100 pet-related products, including automatic...

Instagram’s New Instants App Is a Snapchat Clone for Thirst Traps

Meta launched a new app on Wednesday, called Instants, that integrates with existing Instagram accounts and allows users to...

Must read

Top Home Chef Promo Codes for May 2026

Out of the dozens of services I’ve tested, Home...

Sealy Promo Codes: $100 Off

Sealy is a mattress brand that is tried and...
- Advertisement -

You might also likeRELATED
Recommended to you