Bigscience/bloom
OPT: Open Pre-trained Transformer Language Models
T0: Multitask Prompted Training Enables Zero-Shot Task Generalization
GPT-J-6B: 6B JAX-Based Transformer
https://github.com/HazyResearch/ama_prompting
Scaling Laws for Neural Language Models
Chinchilla: Training Compute-Optimal Large Language Models
Emergent Abilities of Large Language Models
EleutherAI