-
GPT-3: Language Models are Few-Shot Learners
-
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
-
REALM: Retrieval-Augmented Language Model Pre-Training
-
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
-
‘end-to-end’ directory
-
https://github.com/IBM/kgi-slot-filling/tree/re2g
-