PulseAugur
LIVE 07:16:20
research · [19 sources] ·
0
research

New research explores methods to prevent catastrophic forgetting in AI models

Multiple research papers submitted on May 6, 2026, explore novel approaches to continual learning across various AI domains. One paper introduces a replay-based strategy for physics-informed neural operators to mitigate catastrophic forgetting. Another proposes "skill neologisms" using soft tokens to extend LLM capabilities without weight updates. Additionally, research on LLM systems presents a multi-timescale memory dynamics approach for continual knowledge updating, inspired by biological memory. AI

Summary written by gemini-2.5-flash-lite from 19 sources. How we write summaries →

IMPACT These papers explore methods to improve AI's ability to learn continuously without forgetting past knowledge, crucial for adaptive and evolving systems.

RANK_REASON Multiple arXiv papers published on May 6, 2026, detailing new research in continual learning.

Read on arXiv cs.AI →

COVERAGE [19]

  1. arXiv cs.LG TIER_1 · Md Anwar Hossen, Fatema Siddika, Juan Pablo Munoz, Tanya Roosta, Ali Jannesari ·

    CRAFT: Forgetting-Aware Intervention-Based Adaptation for Continual Learning

    arXiv:2605.05732v1 Announce Type: new Abstract: Large language models (LLMs) can acquire new capabilities through fine-tuning, but continual adaptation often leads to catastrophic forgetting. We propose CRAFT, a continual learning framework that avoids updating model weights by i…

  2. arXiv cs.LG TIER_1 · Yazheng Liu, Yuxuan Wan, Rui Xu, Xi Zhang, Sihong Xie, Hui Xiong ·

    Attribution-Guided Continual Learning for Large Language Models

    arXiv:2605.05285v1 Announce Type: new Abstract: Large language models (LLMs) often suffer from catastrophic forgetting in continual learning: after learning new tasks sequentially, they perform worse on earlier tasks. Existing methods mitigate catastrophic forgetting by data repl…

  3. arXiv cs.LG TIER_1 · Yizheng Wang, Mohammad Sadegh Eshaghi, Xiaoying Zhuang, Timon Rabczuk, Yinghua Liu ·

    Replay-Based Continual Learning for Physics-Informed Neural Operators

    arXiv:2605.04832v1 Announce Type: new Abstract: Neural operators generally demonstrate strong predictive performance on in-distribution (ID) problems. However, a critical limitation of existing methods is their significant performance degradation when encountering out-of-distribu…

  4. arXiv cs.AI TIER_1 · Bing Han, Feifei Zhao, Wenxuan Pan, Zhuoya Zhao, Xianqi Li, Qingqun Kong, Yi Zeng ·

    Adaptive Reorganization of Neural Pathways for Continual Learning with Spiking Neural Networks

    arXiv:2309.09550v4 Announce Type: replace-cross Abstract: The human brain can self-organize rich and diverse sparse neural pathways to incrementally master hundreds of cognitive tasks. However, most existing continual learning algorithms for deep artificial and spiking neural net…

  5. arXiv cs.LG TIER_1 · Elvin Hajizada, Danielle Rager, Timothy Shea, Leobardo Campos-Macias, Andreas Wild, Eyke H\"ullermeier, Yulia Sandamirskaya, Mike Davies ·

    Online Continual Learning on Intel Loihi 2 via a Co-designed Spiking Neural Network

    arXiv:2511.01553v2 Announce Type: replace Abstract: AI systems on edge devices require online continual learning -- adapting to non-stationary streams and unfamiliar classes without catastrophic forgetting -- under strict power constraints. We present CLP-SNN, a spiking neural ne…

  6. arXiv cs.LG TIER_1 · Andreas Pattichis, Constantine Dovrolis ·

    Continual Knowledge Updating in LLM Systems: Learning Through Multi-Timescale Memory Dynamics

    arXiv:2605.05097v1 Announce Type: new Abstract: LLMs are trained once, then deployed into a world that never stops changing. External memory compensates for this, but most systems manage it explicitly rather than letting it adapt on its own. Biological memory works differently: c…

  7. arXiv cs.LG TIER_1 · Antonin Berthon, Nicolas Astorga, Mihaela van der Schaar ·

    Skill Neologisms: Towards Skill-based Continual Learning

    arXiv:2605.04970v1 Announce Type: new Abstract: Modern LLMs show mastery over an ever-growing range of skills, as well as the ability to compose them flexibly. However, extending model capabilities to new skills in a scalable manner is an open-problem: fine-tuning and parameter-e…

  8. arXiv cs.CL TIER_1 · Constantine Dovrolis ·

    Continual Knowledge Updating in LLM Systems: Learning Through Multi-Timescale Memory Dynamics

    LLMs are trained once, then deployed into a world that never stops changing. External memory compensates for this, but most systems manage it explicitly rather than letting it adapt on its own. Biological memory works differently: coupled multi-timescale dynamics make new associa…

  9. arXiv cs.AI TIER_1 · Mihaela van der Schaar ·

    Skill Neologisms: Towards Skill-based Continual Learning

    Modern LLMs show mastery over an ever-growing range of skills, as well as the ability to compose them flexibly. However, extending model capabilities to new skills in a scalable manner is an open-problem: fine-tuning and parameter-efficient variants risk catastrophic forgetting, …

  10. arXiv cs.LG TIER_1 · Yinghua Liu ·

    Replay-Based Continual Learning for Physics-Informed Neural Operators

    Neural operators generally demonstrate strong predictive performance on in-distribution (ID) problems. However, a critical limitation of existing methods is their significant performance degradation when encountering out-of-distribution (OOD) data. To address this issue, this wor…

  11. arXiv cs.LG TIER_1 · Ryan King, Gang Li, Bobak Mortazavi, Tianbao Yang ·

    Memory-Efficient Continual Learning with CLIP Models

    arXiv:2605.03866v1 Announce Type: new Abstract: Contrastive Language-Image Pretraining (CLIP) models excel at understanding image-text relationships but struggle with adapting to new data without forgetting prior knowledge. To address this, models are typically fine-tuned using b…

  12. arXiv cs.LG TIER_1 · Chengcheng Xie ·

    Adaptive Data Compression and Reconstruction for Memory-Bounded EEG Continual Learning

    arXiv:2605.03085v1 Announce Type: new Abstract: Electroencephalography (EEG) signals provide millisecond-level temporal resolution but their analysis is limited by remarkable noise and inter-subject variability, making robust personalization difficult under limited annotations. U…

  13. arXiv cs.LG TIER_1 · Tianbao Yang ·

    Memory-Efficient Continual Learning with CLIP Models

    Contrastive Language-Image Pretraining (CLIP) models excel at understanding image-text relationships but struggle with adapting to new data without forgetting prior knowledge. To address this, models are typically fine-tuned using both new task data and a memory buffer of past ta…

  14. arXiv cs.LG TIER_1 · Steven Tang, Xinze Xiong, Anna Hakhverdyan, Andrew Patterson, Jacob Adkins, Jiamin He, Esraa Elelimy, Parham Mohammad Panahi, Martha White, Adam White ·

    Forager: a lightweight testbed for continual learning with partial observability in RL

    arXiv:2605.01131v1 Announce Type: new Abstract: In continual reinforcement learning (CRL), good performance requires never-ending learning, acting, and exploration in a big, partially observable world. Most CRL experiments have focused on loss of plasticity -- the inability to ke…

  15. arXiv cs.LG TIER_1 · Joern Hentsch ·

    MPCS: Neuroplastic Continual Learning via Multi-Component Plasticity and Topology-Aware EWC

    arXiv:2605.02509v1 Announce Type: new Abstract: Continual learning systems face a fundamental tension between plasticity -- acquiring new knowledge -- and stability -- retaining prior knowledge. We introduce MPCS (Multi-Plasticity Continual System), a neuroplastic architecture th…

  16. arXiv cs.LG TIER_1 · Joern Hentsch ·

    MPCS: Neuroplastic Continual Learning via Multi-Component Plasticity and Topology-Aware EWC

    Continual learning systems face a fundamental tension between plasticity -- acquiring new knowledge -- and stability -- retaining prior knowledge. We introduce MPCS (Multi-Plasticity Continual System), a neuroplastic architecture that integrates eleven complementary mechanisms: t…

  17. arXiv cs.LG TIER_1 · Qingyao Ai, Yichen Tang, Changyue Wang, Jianming Long, Weihang Su, Yiqun Liu ·

    MemoryBench: A Benchmark for Memory and Continual Learning in LLM Systems

    arXiv:2510.17281v5 Announce Type: replace Abstract: Scaling up data, parameters, and test-time computation has been the mainstream methods to improve LLM systems (LLMsys), but their upper bounds are almost reached due to the gradual depletion of high-quality data and marginal gai…

  18. arXiv cs.AI TIER_1 · Qisheng Hu, Quanyu Long, Wenya Wang ·

    When Continual Learning Moves to Memory: A Study of Experience Reuse in LLM Agents

    arXiv:2604.27003v1 Announce Type: cross Abstract: Memory-augmented LLM agents offer an appealing shortcut to continual learning: rather than updating model parameters, they accumulate experience in external memory, seemingly sidestepping the stability-plasticity dilemma of parame…

  19. arXiv cs.CV TIER_1 · Shengqin Jiang, Tianqi Kong, Yuankai Qi, Haokui Zhang, Lina Yao, Quan Z. Sheng, Qingshan Liu, Ming-Hsuan Yang ·

    Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning

    arXiv:2511.12090v2 Announce Type: replace Abstract: Prompt-based continual learning methods fine-tune only a small set of additional learnable parameters while keeping the pre-trained model's parameters frozen. It enables efficient adaptation to new tasks while mitigating the ris…