Bibliography (12):

  1. https://x.com/Killa_ru/status/1565740292342484994

  2. https://x.com/Scobleizer/status/1565755925809360896

  3. https://x.com/Scobleizer/status/1560843951287898112

  4. Language Models are Unsupervised Multitask Learners

  5. https://x.com/Scobleizer/status/1589870120780042241

  6. https://x.com/sama/status/1590416386765254656

  7. https://x.com/sama/status/1592622522495045632

  8. Sparse is Enough in Scaling Transformers

  9. Context on the NVIDIA ChatGPT opportunity—and ramifications of large language model enthusiasm

  10. Generating Diverse High-Fidelity Images with VQ-VAE-2