Gradient Episodic Memory for Continual Learning: Difference between revisions

From statwiki
Jump to navigation Jump to search
(Created page with "== Group Member == Yu Xuan Lee, Tsen Yee Heng == Background and Introduction == Supervised learning consist of a training set <math>D_tx={(x_i,y_i)}^n_{i=1}</math>, where <...")
 
No edit summary
Line 4: Line 4:


==  Background and Introduction ==
==  Background and Introduction ==
Supervised learning consist of a training set <math>D_tx={(x_i,y_i)}^n_{i=1}</math>, where <math>x_i\inX</math> and <math>y_i/inY</math>.  
Supervised learning consist of a training set <math>D_{tx}={(x_i,y_i)}^n_{i=1}</math>, where <math>x_i\inX</math> and <math>y_i/inY</math>.  
Gradient Episodic Memory (GEM) is a continual learning model that alleviates forgetting on previous acquired knowledge, while solving new problems more efficiently.
Gradient Episodic Memory (GEM) is a continual learning model that alleviates forgetting on previous acquired knowledge, while solving new problems more efficiently.

Revision as of 00:29, 17 November 2018

Group Member

Yu Xuan Lee, Tsen Yee Heng

Background and Introduction

Supervised learning consist of a training set [math]\displaystyle{ D_{tx}={(x_i,y_i)}^n_{i=1} }[/math], where [math]\displaystyle{ x_i\inX }[/math] and [math]\displaystyle{ y_i/inY }[/math]. Gradient Episodic Memory (GEM) is a continual learning model that alleviates forgetting on previous acquired knowledge, while solving new problems more efficiently.