DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer