504

Hypothesis-driven Stream Learning with Augmented Memory

IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2021
Abstract

Stream learning refers to the ability to acquire and transfer knowledge across a continuous stream of data without forgetting and without repeated passes over the data. A common way to avoid catastrophic forgetting is to intersperse new examples with replays of old examples stored as image pixels or reproduced by generative models. Here, we consider stream learning in image classification tasks and propose a novel hypotheses-driven Augmented Memory Network, which efficiently consolidates previous knowledge with a limited number of hypotheses in the augmented memory and replays relevant hypotheses to avoid catastrophic forgetting. The advantages of hypothesis-driven replay over pixel-level replay and generative replay are two-fold. First, hypothesis-based knowledge consolidation avoids redundant information in the image pixel space and makes memory usage more efficient. Second, hypotheses in the augmented memory can be re-used for learning new tasks, improving generalization and transfer learning ability. We evaluated our method on three stream learning object recognition datasets. Our method performs comparably well or better than state-of-the-art methods, while offering more efficient memory usage. All source code and data are publicly available https://github.com/kreimanlab/AugMem.

View on arXiv
Comments on this paper