GPT-3: Language Models are Few-Shot Learners
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
REALM: Retrieval-Augmented Language Model Pre-Training
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
‘end-to-end’ directory
https://github.com/IBM/kgi-slot-filling/tree/re2g