Bibliography (5):

  1. GPT-3: Language Models are Few-Shot Learners

  2. Generalization without systematicity: On the compositional skills of sequence-to-sequence recurrent networks

  3. https://x.com/denny_zhou/status/1532104072353808384

  4. PaLM: Scaling Language Modeling with Pathways

  5. DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs