AI Enters Kindergarten to Master Complex Tasks

New York University

We need to learn our letters before we can learn to read and our numbers before we can learn how to add and subtract. The same principles are true with AI, a team of New York University scientists has shown through laboratory experiments and computational modeling.

In their work, published in the journal Nature Machine Intelligence, researchers found that when recurrent neural networks (RNNs) are first trained on simple cognitive tasks, they are better equipped to handle more difficult and complex ones later on.

The paper's authors labeled this form of training kindergarten curriculum learning as it centers on first instilling an understanding of basic tasks and then combining knowledge of these tasks in carrying out more challenging ones.

"From very early on in life, we develop a set of basic skills like maintaining balance or playing with a ball," explains Cristina Savin, an associate professor in NYU's Center for Neural Science and Center for Data Science. "With experience, these basic skills can be combined to support complex behavior—for instance, juggling several balls while riding a bicycle.

"Our work adopts these same principles in enhancing the capabilities of RNNs, which first learn a series of easy tasks, store this knowledge, and then apply a combination of these learned tasks to successfully complete more sophisticated ones."

RNNs—neural networks that are designed to process sequential information based on stored knowledge—are particularly useful in speech recognition and language translation. However, when it comes to complex cognitive tasks, training RNNs with existing methods can prove difficult and fall short of capturing crucial aspects of animal and human behavior that AI systems aim to replicate.

To address this, the study's authors—who also included David Hocker, a postdoctoral researcher in NYU's Center for Data Science, and Christine Constantinople, a professor in NYU's Center for Data Science—first conducted a series of experiments with laboratory rats.

The animals were trained to seek out a water source in a box with several compartmentalized ports. However, in order to know when and where the water would be available, the rats needed to learn that delivery of the water was associated with certain sounds and the illumination of the port's lights—and that the water was not delivered immediately after these cues. In order to reach the water, then, the animals needed to develop basic knowledge of multiple phenomena (e.g., sounds precede water delivery, waiting after the visual and audio cues before trying to access the water) and then learn to combine these simple tasks in order to complete a goal (water retrieval).

These results pointed to principles of how the animals applied knowledge of simple tasks in undertaking more complex ones.

The scientists took these findings to train RNNs in a similar fashion—but, instead of water retrieval, the RNNs managed a wagering task that required these networks to build upon basic decision making in order to maximize the payoff over time. They then compared this kindergarten curriculum learning approach to existing RNN-training methods.

Overall, the team's results showed that the RNNs trained on the kindergarten model learned faster than those trained on current methods.

"AI agents first need to go through kindergarten to later be able to better learn complex tasks," observes Savin. "Overall, these results point to ways to improve learning in AI systems and call for developing a more holistic understanding of how past experiences influence learning of new skills."

This research was funded by grants from the National Institute of Mental Health (1R01MH125571-01, 1K01MH132043-01A1) and conducted using research computing resources of the Empire AI consortium, with support from the State of New York, the Simons Foundation, and the Secunda Family Foundation.

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.