Memorization Without Overfitting: Analyzing the Training Dynamics of Large Language Models
Continual Learning with Foundation Models: An Empirical Study of Latent Replay
Donβt Stop Learning: Towards Continual Learning for the CLIP Model
Lifelong Pretraining: Continually Adapting Language Models to Emerging Corpora
DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning
T0: Multitask Prompted Training Enables Zero-Shot Task Generalization
Effect of scale on catastrophic forgetting in neural networks
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Continual Pre-Training Mitigates Forgetting in Language and Vision
An Empirical Investigation of the Role of Pre-training in Lifelong Learning