Bibliography:

  1. https://openai.com/blog/chatgpt/

  2. https://www.reddit.com/r/mlscaling/comments/1g9r65z/introducing_computer_use_a_new_claude_35_sonnet/lt89drx/

  3. https://www.reddit.com/r/slatestarcodex/comments/1gsv897/gwern_on_the_diminishing_returns_to_scaling_and/

  4. https://www.reddit.com/r/mlscaling/comments/1gswayg/gwern_on_the_diminishing_returns_to_scaling_and/

  5. Machine Learning Scaling

  6. https://openai.com/index/gpt-4-research/

  7. Language Models are Unsupervised Multitask Learners

  8. GPT-3: Language Models are Few-Shot Learners

  9. GPQA: A Graduate-Level Google-Proof Q&A Benchmark

  10. Chinchilla: Training Compute-Optimal Large Language Models

  11. $2024

  12. Scaling Laws for Neural Language Models