Publication:
Adaptive Neural Networks for Online Domain Incremental Continual Learning

Loading...
Thumbnail Image

Publisher link

Rights

This is an author’s accepted version of a conference paper published in Proceedings of 25th International Conference on Discovery Science (DS 2022), LNAI 13601. © 2022 Springer.

Abstract

Continual Learning (CL) poses a significant challenge to Neural Network (NN)s, where the data distribution changes from one task to another. In Online domain incremental continual learning (OD-ICL), this distribution change happens in the input space without affecting the label distribution. In order to adapt to such changes, the model being trained risks forgetting previously learned knowledge (stability). On the other hand, enforcing that the model preserves past knowledge will cause it to fail to learn new concepts (plasticity). We propose Online Domain Incremental Networks (ODIN), a novel method to alleviate catastrophic forgetting by automatically detecting the end of a task using concept drift detection. As a consequence, ODIN does not require the specification of task ids. ODIN maintains a pool of NNs, each trained on a single task and frozen for further updates. A Task Predictor (TP) is trained to select the most suitable NN from the frozen pool for prediction. We compare ODIN against popular regularization and replay methods. It outperforms regularization methods and achieves comparable predictive performance to replay methods.

Citation

Series name

Date

Publisher

Springer Nature

Degree

Type of thesis

Supervisor

DOI

Link to supplementary material

Research Projects

Organizational Units

Journal Issue