AI can better retain what it learns by mimicking human sleep

Human brains consolidate memories while sleeping. Could AI systems use the same technique?

Giulia Fiori Photography/Getty Images

Building AIs that sleep and dream can lead to better results and more reliable models, according to researchers who aim to replicate the architecture and behaviour of the human brain. But other experts say recreating the intelligence we see within ourselves may not be the most fruitful path for AI research.

Concetto Spampinato and his colleagues at the University of Catania, Italy, were looking for ways to avoid a phenomenon known as “catastrophic forgetting”, where an AI model trained to do a new task loses the ability to carry out jobs it previously aced. For instance, a model trained to identify animals could learn to spot different fish species, but then it might inadvertently lose its proficiency at recognising birds.

They developed a new method of training AI called wake-sleep consolidated learning (WSCL), which mimics the way human brains reinforce new information. People shuffle short-term memories of experiences and lessons learned throughout the day into long-term memories while sleeping. The researchers say this method of learning can be applied to any existing AI.

Models using WSCL are trained as usual on a set of data for the “awake” phase. But they are also programmed to have periods of “sleeping”, where they parse through a sample of awake data, as well as a highlight reel from previous lessons.

Take an animal identification model more recently trained on images of marine life: during a sleep period, it would be shown snapshots of fishes, but also a smattering of birds, lions and elephants from older lessons. Spampinato says this is akin to humans mulling over new and old memories while sleeping, spotting connections and patterns and integrating them into our minds. The new data teaches the AI a fresh ability, while the remainder of the old data prevents the recently acquired skill from pushing out existing ones.

Crucially, WSCL also has a period of “dreaming”, when it consumes entirely novel data made from mashing together previous concepts. For instance, the animal model might be fed abstract images showing combinations of giraffes crossed with fish, or lions crossed with elephants. Spampinato says this phase helps to merge previous paths of digital “neurons”, freeing up space for other concepts in the future. It also primes unused neurons with patterns that will help them pick up new lessons more easily.

“The point is that, as you gather new knowledge, you just combine more complex patterns,” says Spampinato. “You might combine panda, giraffe and leopard, so you’ll create this very strange mythological figure, and that will force the model to learn more complex patterns that maybe in the future can be reused.”

Spampinato tested three existing AI models using a traditional training method, followed by WSCL training. Then he and his team compared the performances using three standard benchmarks for image identification. The researchers found their newly developed technique led to a significant accuracy boost – the sleep-trained models were 2 to 12 per cent more likely to correctly identify the contents of an image. They also measured an increase in the WSCL systems’ “forward transfer”, a metric indicating how much old knowledge a model uses to learn a new task. The research indicated AI trained with the sleep method remembered old tasks better than the traditionally trained systems.

Despite these promising results, some experts say using the human brain as a blueprint is not necessarily the best way to boost AI performance. Andrew Rogoyski at the University of Surrey in the UK says AI research is still in its infancy, and designing new models is often akin to alchemy. He also warns against anthropomorphising AI architecture or rigidly mimicking the human brain in ever greater detail.

“The human brain should not be regarded as the ultimate architecture for intelligence,” he says. “It’s the result of millions of years of evolution and an unimaginably wide range of stimuli. We may develop AIs that have structures completely different from their biological designers.”

Still, Rogoyski says the sleep training method seems interesting. Instead of having the dreaming AI mimic human beings, however, he suggests another biological inspiration: dolphins, which have the ability to “sleep” with one part of the brain while another remains alert, switching as needed. After all, an AI that requires hours of sleep is not ideal for commercial applications.

Topics:

[colabot7]

Leave a Comment

Your email address will not be published. Required fields are marked *