T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Challenging BIG-Bench Tasks (BBH) and Whether Chain-of-Thought Can Solve Them
TyDiQA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages
https://github.com/google-research/t5x/blob/main/docs/models.md#flan-t5-checkpoints
Self-Consistency Improves Chain-of-Thought Reasoning in Language Models
https://prod.hypermind.com/ngdp/en/showcase2/showcase.html?sc=JSAI
https://www.metaculus.com/questions/11676/mmlu-sota-in-2023-2025/
T0: Multitask Prompted Training Enables Zero-Shot Task Generalization
Tk-Instruct: Benchmarking Generalization via In-Context Instructions on 1,600+ Language Tasks
ByT5: Towards a token-free future with pre-trained byte-to-byte models