TY - JOUR
T1 - Lifelong Language Learning with the Most Forgotten Knowledge
AU - Choi, Heejeong
AU - Kang, Pilsung
N1 - Funding Information:
This work was supported in part by the National Research Foundation of Korea (NRF) funded by the Korean Government (MSIT) under Grant NRF-2019R1F1A1060338 and Grant NRF-2019R1A4A1024732.
Publisher Copyright:
© 2013 IEEE.
PY - 2021
Y1 - 2021
N2 - Lifelong language learning enables a language model to accumulate knowledge through training on a stream of text data. Recent research on lifelong language learning is based on samples of previous tasks from an episodic memory or generative model. LAMOL, a representative generative model-based lifelong language learning model, preserves the previous information with the generated pseudo-old samples, which are suboptimal. In this paper, we propose an improved version of LAMOL, MFK-LAMOL, which constructs a generative replay using a more effective method. When a new task is received, MFK-LAMOL replays sufficient previous data and retrieves important examples for training alongside the new task. Specifically, it selects the examples with the most forgotten knowledge learned from previous tasks based on the extent to which they include knowledge that has been forgotten after learning new information. We showed that the proposed method outperforms LAMOL on a stream of three different natural language processing tasks.
AB - Lifelong language learning enables a language model to accumulate knowledge through training on a stream of text data. Recent research on lifelong language learning is based on samples of previous tasks from an episodic memory or generative model. LAMOL, a representative generative model-based lifelong language learning model, preserves the previous information with the generated pseudo-old samples, which are suboptimal. In this paper, we propose an improved version of LAMOL, MFK-LAMOL, which constructs a generative replay using a more effective method. When a new task is received, MFK-LAMOL replays sufficient previous data and retrieves important examples for training alongside the new task. Specifically, it selects the examples with the most forgotten knowledge learned from previous tasks based on the extent to which they include knowledge that has been forgotten after learning new information. We showed that the proposed method outperforms LAMOL on a stream of three different natural language processing tasks.
KW - Lifelong language learning
KW - a stream of text data
KW - catastrophic forgetting
KW - generative replay
KW - natural language processing
UR - http://www.scopus.com/inward/record.url?scp=85104201655&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2021.3071787
DO - 10.1109/ACCESS.2021.3071787
M3 - Article
AN - SCOPUS:85104201655
SN - 2169-3536
VL - 9
SP - 57941
EP - 57948
JO - IEEE Access
JF - IEEE Access
M1 - 9399079
ER -