DeepMind’s virtual playground suggests path to general AI

XLand marks the spot.
Sign up for the Freethink Weekly newsletter!
A collection of our favorite stories straight to your inbox

DeepMind has created a virtual playground that shows a path to creating general AI — the holy grail of artificial intelligence. 

Reinforcement learning: If you want to train an AI to play chess, you can set up a virtual chessboard, list the rules, and let the AI learn the game through trial and error.

When it does something “right,” such as capturing a pawn, you give it a reward. When it does something majorly right, like winning the game, you give it a bigger reward.

Eventually, the AI will learn what it needs to do to get the most rewards, and boom, you have an AI that can beat any human at chess.

“This marks an important step toward creating more general agents.”

DeepMind

The challenge: This process is called reinforcement learning, and it’s one of the most effective ways to train AIs. However, it has a major limitation: at the end of the training, the AI only knows how to do one specific thing.

Even trying to train an AI that knows how to do that one thing (chess) to do something similar (such as Shogi, aka Japanese chess) requires starting the reinforcement learning process from scratch.

General AI: It would be useful to have a general AI that could use its smarts to solve all sorts of problems, including ones it has never seen before, just like humans do.

General AI doesn’t currently exist, though, because no one has figured out how to teach a machine to succeed at tasks it wasn’t specifically trained on.

In theory, we could just train an agent on everything, one task at a time, but that would require so much training data and time that it’s simply not feasible.

Welcome to XLand: Google sister company DeepMind has now highlighted a potential path to general AI.

It designed a virtual world called ​​“XLand,” where AI agents could navigate environments that look a bit like Battle Courses from Mario Kart. It then built an algorithm that could create billions of different game-like tasks for the AIs to complete in XLand. 

The agents were rewarded for correctly completing tasks, just like they would in a standard reinforcement learning environment, and each new task was designed to be just hard enough to keep the agent learning something new.

The results: By the end of the study, the AIs were able to complete a range of tasks and could rapidly master games that completely stumped new AIs trained from scratch. 

“We find the agent exhibits general, heuristic behaviours such as experimentation, behaviours that are widely applicable to many tasks rather than specialised to an individual task,” DeepMind wrote in a blog post.

“This new approach marks an important step toward creating more general agents with the flexibility to adapt rapidly within constantly changing environments,” it continued.

The next steps: To be clear, DeepMind’s agents aren’t general AI, but they are more well-rounded problem-solvers than AIs trained using traditional, narrow reinforcement learning.

That means the algorithm-as-taskmaster approach detailed in the researchers’ paper, which still needs to undergo peer-review, might be how we can create the more capable AIs of the future.

We’d love to hear from you! If you have a comment about this article or if you have a tip for a future Freethink story, please email us at [email protected].

Sign up for the Freethink Weekly newsletter!
A collection of our favorite stories straight to your inbox
Related
AI is now designing chips for AI
AI-designed microchips have more power, lower cost, and are changing the tech landscape.
Why futurist Amy Webb sees a “technology supercycle” headed our way
Amy Webb’s data suggests we are on the cusp of a new tech revolution that will reshape the world in much the same way the steam engine and internet did in the past.
AI chatbots may ease the world’s loneliness (if they don’t make it worse)
AI chatbots may have certain advantages when roleplaying as our friends. They may also come with downsides that make our loneliness worse.
Will AI supercharge hacking — if it hasn’t already?
The future of hacking is coming at us fast, and it isn’t clear yet whether AI will help attackers and defenders more.
No, LLMs still can’t reason like humans. This simple test reveals why.
Most AI models are incredible at taking tests but easily bamboozled by basic reasoning. “Simple Bench” shows us why.
Up Next
ar quest
Subscribe to Freethink for more great stories