Bibliography (10):

  1. Bigscience/bloom

  2. OPT: Open Pre-trained Transformer Language Models

  3. T0: Multitask Prompted Training Enables Zero-Shot Task Generalization

  4. GPT-J-6B: 6B JAX-Based Transformer

  5. https://github.com/HazyResearch/ama_prompting

  6. Scaling Laws for Neural Language Models

  7. Chinchilla: Training Compute-Optimal Large Language Models

  8. Emergent Abilities of Large Language Models

  9. EleutherAI

  10. Wikipedia Bibliography:

    1. Entropy (information theory)