Bibliography (4):

  1. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension

  2. T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

  3. https://github.com/belindal/state-probes