Memory-based Parameter Adaptation
04 Jul 2018Introduction
-
Standard Deep Learning networks are not suitable for continual learning setting...
Standard Deep Learning networks are not suitable for continual learning setting...
The paper explores knowledge distillation (KD) from the perspective of transferring...
The paper presents a simple yet effective approach for transferring knowledge...
The paper presents a general message passing architecture called as Message...
Convolutional Neural Networks are extremely good feature extractors in the sense...
Empirical evidence indicates that at training time, the neural networks need...
Conventional wisdom says that when training neural networks, learning rate should...
Information Extraction - Given a query to be answered and an...
Catastrophic Forgetting refers to the phenomenon where when a learning system...