Bibliography (4):
Attention Is All You Need
ETC: Encoding Long and Structured Inputs in Transformers
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer