Elon Musk Is Wrong About AI

In painting a rosy and likely unrealistic picture of what AI can and can't do, Musk has, in our view, misled the public about how far we still have to go.

Image
Binary and technology on a blue background showing AI

Elon Musk has a habit of using Twitter and interviews to make big statements. For instance, Musk told Jack Dorsey via tweet that AGI-artificial general intelligence, or AI with the power and flexibility of human intelligence--would most likely be here by 2029.

And when Elon talks, people listen. But should they?

He has a history of making bold predictions, not always correct; those self-driving taxis he promised still aren't here, for example. In this particular instance, the idea that some quantum jump in AI is imminent might actually cause some people to panic, especially given that Musk himself once famously told a crowd at MIT that, "with artificial intelligence, we are summoning the demon." At the same time, suggesting that humanlike intelligence is not far away might distract from all the current flaws in AI that so desperately need fixing.

The truth is, there is a giant gap between today's AI, which is largely pattern recognition, and the kind of Star Trek-computer-level AI that Musk is dreaming about. Yes, AI can already do some amazing things, such as speech recognition, with the ability to hold surrealistic but entertaining conversations about virtually any topic. But when it comes to reliability, dependability and coherence, current AI is nowhere near what it needs to be. Despite years of promises, AI continues to regularly make bizarre and unexpected errors of "discomprehension." It also perpetuates stereotypes; spreads misinformation; and still fails even at everyday tasks like human-level driving, especially in unexpected circumstances. Recently, a "summoned" Tesla crashed into a $3 million jet that was parked at a mostly empty airport. Inside the field, these kinds of challenges are well-known, but there are no firm fixes at hand.

Remedying AI's current flaws (and using the AI we actually have now wisely) must start with realism. Building an AI that is genuinely trustworthy is one of the most important but also challenging engineering missions of our time. But being glib about it isn't helping. In painting a rosy and likely unrealistic picture, Musk has, in our view, misled the public about how far we still have to go.

With so much at stake, we decided to call BS.

It began when one of us, Gary Marcus, drafted a $100,000 bet. In essence, the bet highlights the disconnection between Musk's latest claims about the future of AI and current reality. In the spirit of serious betting, Marcus laid out five very specific conditions.

To really say that AGI had been achieved, it would have to clear at least three of the following five benchmarks of intelligence, compiled in collaboration with NYU computer scientist Ernest Davis. To be considered artificial general intelligence, AI would need to be able to accomplish some of the following:

  1. Watch movies and tell us accurately what is going on. Who are the characters? What are their conflicts and motivations? Et cetera.
  2. Read novels and reliably answer questions about plot, character, conflicts, motivations, etc. The key is to go beyond the literal text and show a real understanding of the material.
  3. Work as a competent cook in any old random kitchen (a tip of the hat to Steve Wozniak's cup-of-coffee benchmark).
  4. Reliably construct bug-free code of more than 10,000 lines from natural language specification, or by interactions with a nonexpert user. (Gluing together code from existing libraries doesn't count.)
  5. Take arbitrary proofs from the mathematical literature, written in natural language, and convert them into a symbolic form suitable for symbolic verification.

The other of us, Vivek Wadhwa, thought the bet was terrific, fair and provocative, and something that could move the field of AI forward. So Wadhwa decided to match Marcus's wager. Within a couple hours, there was a flurry on Twitter, and Marcus's Substack had close to 10,000 views; soon, other experts in the field also offered their support to $500,000. But not a word from Musk.

See also: The Real Disruption From Robotics, AI

Then writer and futurist Kevin Kelly, who co-founded the Long Now Foundation, offered to host the bet on his website side by side with an earlier and related bet that Ray Kurzweil made with Mitch Kapor. Ben Goertzel, for decades one of the leaders in trying to make AGI into something real, rather than just a fantasy, tweeted that he thought the tests would signify real progress. World Summit AI, the world's leading AI conference, offered to host a debate. Others wondered aloud which benchmarks might fall first, and in what order.

Despite all that excitement in the AI community, there has still been no word from Musk.

Half a million bucks is chump change, of course, for someone who is perhaps the richest person in the world. But it is real money to us, and it symbolizes something important: the value of getting public voices who hype AI's near-term prospects to stand by their claims.

Spreading misinformation about the potential of AI and its likely progress may serve Tesla by diverting attention from the many problems it has with its self-driving software, but it doesn't serve the public. If Musk believes what he says, he should stand up and take the bet; if not, he should own up to the reality that his pronouncements are little more than off-the-cuff hunches that even Musk himself realizes aren't worth the virtual paper he's printed them on.

This article was written by Vivek Wadhwa and Gary Marcus.


Vivek Wadhwa

Profile picture for user VivekWadhwa

Vivek Wadhwa

Vivek Wadhwa is a fellow at Arthur and Toni Rembe Rock Center for Corporate Governance, Stanford University; director of research at the Center for Entrepreneurship and Research Commercialization at the Pratt School of Engineering, Duke University; and distinguished fellow at Singularity University.

MORE FROM THIS AUTHOR

Read More