Publication Type
Conference Proceeding Article
Version
publishedVersion
Publication Date
5-2021
Abstract
Continual learning methods with fixed architectures rely on a single network to learn models that can perform well on all tasks. As a result, they often only accommodate common features of those tasks but neglect each task's specific features. On the other hand, dynamic architecture methods can have a separate network for each task, but they are too expensive to train and not scalable in practice, especially in online settings. To address this problem, we propose a novel online continual learning method named ``Contextual Transformation Networks” (CTN) to efficiently model the \emph{task-specific features} while enjoying neglectable complexity overhead compared to other fixed architecture methods. Moreover, inspired by the Complementary Learning Systems (CLS) theory, we propose a novel dual memory design and an objective to train CTN that can address both catastrophic forgetting and knowledge transfer simultaneously. Our extensive experiments show that CTN is competitive with a large scale dynamic architecture network and consistently outperforms other fixed architecture methods under the same standard backbone. Our implementation can be found at \url{https://github.com/phquang/Contextual-Transformation-Network}.
Discipline
Databases and Information Systems | OS and Networks
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
International Conference on Learning Representations, ICLR 2021, Vienna, Austria, May 4
First Page
1
Last Page
20
City or Country
Vienna, Austria
Citation
PHAM, Quang; LIU, Chenghao; SAHOO, Doyen; and HOI, Steve C. H..
Contextual transformation networks for online continual learning. (2021). International Conference on Learning Representations, ICLR 2021, Vienna, Austria, May 4. 1-20.
Available at: https://ink.library.smu.edu.sg/sis_research/10171
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://openreview.net/forum?id=zx_uX-BO7CH