Bibliography (4):

  1. Attention Is All You Need

  2. ETC: Encoding Long and Structured Inputs in Transformers

  3. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization

  4. T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer