[ContinualAI Reading Group]: "Small-Task Incremental Learning"

[May 22th 2020] ContinualAI Reading Group: "Small-Task Incremental Learning"

Abstract: Lifelong learning has attracted much attention, but existing works still struggle to fight catastrophic forgetting and accumulate knowledge over long stretches of incremental learning. In this work, we propose PODNet, a model inspired by representation learning. By carefully balancing the compromise between remembering the old classes and learning new ones, PODNet fights catastrophic forgetting, even over very long runs of small incremental tasks-- a setting so far unexplored by current works. PODNet innovates on existing art with an efficient spatial-based distillation-loss applied throughout the model and a representation comprising multiple proxy vectors for each class. We validate those innovations thoroughly, comparing PODNet with three state-of-the-art models on three datasets: CIFAR100, ImageNet100, and ImageNet1000. Our results showcase a significant advantage of PODNet over existing art, with accuracy gains of 12.10, 4.83, and 2.85 percentage points, respectively.

Arthur Douillard

:round_pushpin: Youtube recording : https://www.youtube.com/watch?v=9xm4P4Kss54
:round_pushpin: Paper pre-print : https://arxiv.org/abs/2004.13513
:round_pushpin: Slides: https://arthurdouillard.com/files/talks/douillard2020podnet_continualai.pdf
:round_pushpin: Code: https://github.com/arthurdouillard/incremental_learning.pytorch

1 Like