At first glance, a recent batch of research papers produced by a prominent artificial intelligence lab at the University of British Columbia in Vancouver might not seem that notable. Featuring incremental improvements on existing algorithms and ideas, they read like the contents of a middling AI conference or journal.
But the research is, in fact, remarkable. Thatâs because itâs entirely the work of an âAI scientistâ developed at the UBC lab together with researchers from the University of Oxford and a startup called Sakana AI.
The project demonstrates an early step toward what might prove a revolutionary trick: letting AI learn by inventing and exploring novel ideas. Theyâre just not super novel at the moment. Several papers describe tweaks for improving an image-generating technique known as diffusion modeling; another outlines an approach for speeding up learning in deep neural networks.
âThese are not breakthrough ideas. Theyâre not wildly creative,â admits Jeff Clune, the professor who leads the UBC lab. âBut they seem like pretty cool ideas that somebody might try.â
As amazing as todayâs AI programs can be, they are limited by their need to consume human-generated training data. If AI programs can instead learn in an open-ended fashion, by experimenting and exploring âinterestingâ ideas, they might unlock capabilities that extend beyond anything humans have shown them.
Cluneâs lab had previously developed AI programs designed to learn in this way. For example, one program called Omni tried to generate the behavior of virtual characters in several video-game-like environments, filing away the ones that seemed interesting and then iterating on them with new designs. These programs had previously required hand-coded instructions in order to define interestingness. Large language models, however, provide a way to let these programs identify whatâs most intriguing. Another recent project from Cluneâs lab used this approach to let AI programs dream up the code that allows virtual characters to do all sorts of things within a Roblox-like world.
The AI scientist is one example of Cluneâs lab riffing on the possibilities. The program comes up with machine learning experiments, decides what seems most promising with the help of an LLM, then writes and runs the necessary codeârinse and repeat. Despite the underwhelming results, Clune says open-ended learning programs, as with language models themselves, could become much more capable as the computer power feeding them is ramped up.
âIt feels like exploring a new continent or a new planet,â Clune says of the possibilities unlocked by LLMs. âWe don’t know what we’re going to discover, but everywhere we turn, there’s something new.â
Tom Hope, an assistant professor at the Hebrew University of Jerusalem and a research scientist at the Allen Institute for AI (AI2), says the AI scientist, like LLMs, appears to be highly derivative and cannot be considered reliable. âNone of the components are trustworthy right now,â he says.
Hope points out that efforts to automate elements of scientific discovery stretch back decades to the work of AI pioneers Allen Newell and Herbert Simon in the 1970s, and, later, the work of Pat
Langley at the Institute for the Study of Learning and Expertise. He also notes that several other research groups, including a team at AI2, have recently harnessed LLMs to help with generating hypotheses, writing papers, and reviewing research. âThey captured the zeitgeist,â Hope says of the UBC team. âThe direction is, of course, incredibly valuable, potentially.â
Whether LLM-based systems can ever come up with truly novel or breakthrough ideas also remains unclear. âThatâs the trillion-dollar question,â Clune says.
Even without scientific breakthroughs, open-ended learning may be vital to developing more capable and useful AI systems in the here and now. A report posted this month by Air Street Capital, an investment firm, highlights the potential of Cluneâs work to develop more powerful and reliable AI agents, or programs that autonomously perform useful tasks on computers. The big AI companies all seem to view agents as the next big thing.
This week, Cluneâs lab revealed its latest open-ended learning project: an AI program that invents and builds AI agents. The AI-designed agents outperform human-designed agents in some tasks, such as math and reading comprehension. The next step will be devising ways to prevent such a system from generating agents that misbehave. âIt’s potentially dangerous,â Clune says of this work. âWe need to get it right, but I think it’s possible.â