-
https://github.com/socialfoundations/tttlm
-
Large-Scale Retrieval for Reinforcement Learning
-
In-Context Pretraining (ICP): Language Modeling Beyond Document Boundaries
-
โ GPT-3 Creative Fiction ยง Prompts As Programming
-
https://github.com/EleutherAI/The-Pile
-
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
-
EleutherAI/gpt-Neo: An Implementation of Model Parallel GPT-2 and GPT-3-Style Models Using the Mesh-Tensorflow Library.
-
Recurrent Neural Network Based Language Model ยง Dynamic Evaluation
-
Language Models are Unsupervised Multitask Learners
-
2023-hardt-figure6-perplexitiesdecreasewhentrainingonincreasinglymoreneighborsusinggpt2onthepile.jpg
-