Cracking the Code of Memory Loss in AI

Memories can be as tricky to hold onto for machines as they can be for humans. To help understand why artificial agents develop holes in their cognitive processes, electrical engineers at Ohio State University have analyzed how much a “continual learning” process impacts their overall performance. 

Study: Future AI algorithms have potential to learn like humans, say researchers. Image Credit: 3rdtimeluckystudio / ShutterstockStudy: Future AI algorithms have potential to learn like humans, say researchers. Image Credit: 3rdtimeluckystudio / Shutterstock

Continual learning is when a computer is trained to continuously learn a sequence of tasks, using its accumulated knowledge from old tasks to better understand new tasks. 

Yet one foremost hurdle scientists still need to overcome to achieve such heights is learning how to circumvent the machine learning equivalent of memory loss – a process which in AI agents is known as “catastrophic forgetting.” As artificial neural networks are trained on one new task after another, they tend to lose the information gained from those previous tasks, an issue that could become problematic as society comes to rely on AI systems more and more, said Ness Shroff, an Ohio Eminent Scholar and professor of computer science and engineering at The Ohio State University.

“As automated driving applications or other robotic systems are taught new things, it’s important that they don’t forget the lessons they’ve already learned for our safety and theirs,” said Shroff. “Our research delves into the complexities of continuous learning in these artificial neural networks, and what we found are insights that begin to bridge the gap between how a machine learns and how a human learns.”

Researchers found that in the same way that people might struggle to recall contrasting facts about similar scenarios but remember inherently different situations with ease, artificial neural networks can recall information better when faced with diverse tasks in succession, instead of ones that share similar features, Shroff said. 

The team, including Ohio State postdoctoral researchers Sen Lin and Peizhong Ju and professors Yingbin Liang and Shroff, will present their research this month at the 40th annual International Conference on Machine Learning in Honolulu, Hawaii, a flagship conference in machine learning. 

While it can be challenging to teach autonomous systems to exhibit this kind of dynamic, lifelong learning, possessing such capabilities would allow scientists to scale up machine learning algorithms at a faster rate and easily adapt them to handle evolving environments and unexpected situations. Essentially, the goal for these systems would be for them to one day mimic the learning capabilities of humans.

Traditional machine learning algorithms are trained on data all at once, but this team’s findings showed that factors like task similarity, negative and positive correlations, and even the order in which an algorithm is taught a task matter in the length of time an artificial network retains certain knowledge. 

For instance, to optimize an algorithm’s memory, said Shroff, dissimilar tasks should be taught early on in the continual learning process. This method expands the network’s capacity for new information and improves its ability to subsequently learn more similar tasks. 

Shroff said their work is particularly important as understanding the similarities between machines and the human brain could lead to a deeper understanding of AI. 

“Our work heralds a new era of intelligent machines that can learn and adapt like their human counterparts,” he said. 

The study was supported by the National Science Foundation and the Army Research Office.

 
Source:
Journal reference:
  • Lin, S., Ju, P., Liang, Y., & Shroff, N. (2023). Theory on Forgetting and Generalization of Continual Learning. ArXiv. /abs/2302.05836, https://arxiv.org/abs/2302.05836
 

Article Revisions

  • Jun 24 2024 - Fixed broken link to journal paper - https://arxiv.org/abs/2302.05836

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Boost Machine Learning Trust With HEX's Human-in-the-Loop Explainability