Loading...
Adaptive Neural Networks for Online Domain Incremental Continual Learning
Abstract
Continual Learning (CL) poses a significant challenge to Neural Network (NN)s, where the data distribution changes from one task to another. In Online domain incremental continual learning (OD-ICL), this distribution change happens in the input space without affecting the label distribution. In order to adapt to such changes, the model being trained risks forgetting previously learned knowledge (stability). On the other hand, enforcing that the model preserves past knowledge will cause it to fail to learn new concepts (plasticity). We propose Online Domain Incremental Networks (ODIN), a novel method to alleviate catastrophic forgetting by automatically detecting the end of a task using concept drift detection. As a consequence, ODIN does not require the specification of task ids. ODIN maintains a pool of NNs, each trained on a single task and frozen for further updates. A Task Predictor (TP) is trained to select the most suitable NN from the frozen pool for prediction. We compare ODIN against popular regularization and replay methods. It outperforms regularization methods and achieves comparable predictive performance to replay methods.
Type
Conference Contribution
Type of thesis
Series
Citation
Date
2022
Publisher
Springer Nature
Degree
Supervisors
Rights
This is an author’s accepted version of a conference paper published in Proceedings of 25th International Conference on Discovery Science (DS 2022), LNAI 13601. © 2022 Springer.