Explaining Robustness to Catastrophic Forgetting Through Incremental Concept Formation

Nicki Barari, Edward Kim, Christopher J. MacLellan

Proceedings of the Twelfth Annual Conference on Advances in Cognitive Systems

2025

Abstract

Catastrophic forgetting remains a central challenge in continual learning, where models are required to integrate new knowledge over time without losing what they have previously learned. In prior work, we introduced Cobweb/4V, a hierarchical concept formation model that exhibited robustness to catastrophic forgetting in visual domains. Motivated by this robustness, we examine three hypotheses regarding the factors that contribute to such stability: (1) adaptive structural reorganization enhances knowledge retention, (2) sparse and selective updates reduce interference, and (3) information-theoretic learning based on sufficiency statistics provides advantages over gradient-based backpropagation. To test these hypotheses, we compare Cobweb/4V with neural baselines, including CobwebNN, a neural implementation of the Cobweb framework introduced in this work. Experiments on datasets of varying complexity (MNIST, Fashion-MNIST, MedMNIST, and CIFAR10) show that adaptive restructuring enhances learning plasticity, sparse updates help mitigate interference, and the information-theoretic learning process preserves prior knowledge without revisiting past data. Together, these findings provide insight into mechanisms that can mitigate catastrophic forgetting and highlight the potential of concept-based, information-theoretic approaches for building stable and adaptive continual learning systems.

Topics:Computer VisionConcept LearningCatastrophic Forgetting
Projects:Cobweb

BibTeX

@inproceedings{barari-acs-2025,
  title     = {Explaining Robustness to Catastrophic Forgetting Through Incremental Concept Formation},
  author    = {Barari, Nicki and Kim, Edward and MacLellan, Christopher J.},
  booktitle = {Proceedings of the Twelfth Annual Conference on Advances in Cognitive Systems},
  pages     = {315-332},
  year      = {2025},
}

← All Publications