https://deepmind.google/
https://www.reddit.com/r/mlscaling/comments/vq6qh1/demis_hassabis_gato_is_our_most_general_agent_so/ienfekn/
Decision Transformer: Reinforcement Learning via Sequence Modeling
The Arcade Learning Environment: An Evaluation Platform for General Agents
A domain-specific supercomputer for training deep neural networks
β scaling-hypothesis#blessings-of-scale
[Transclude the forward-link's
context]
The Bitter Lesson
Multi-task Deep Reinforcement Learning with PopArt
https://www.metaculus.com/questions/3479/date-weakly-general-ai-is-publicly-known/
https://www.lesswrong.com/posts/5onEtjNEhqcfX3LXG/a-generalist-agent-new-deepmind-publication
https://news.ycombinator.com/item?id=31355657
https://x.com/ohlennart/status/1524877652867309570
PaLM: Scaling Language Modeling with Pathways
Chinchilla: Training Compute-Optimal Large Language Models
DeepMind Lab
Imagination-Augmented Agents for Deep Reinforcement Learning
BabyAI: A Platform to Study the Sample Efficiency of Grounded Language Learning
https://github.com/google-deepmind/dm_control
Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning
Procgen Benchmark: Weβre releasing Procgen Benchmark, 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizable skills
Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
One Policy to Control Them All: Shared Modular Policies for Agent-Agnostic Control
Offline Learning from Demonstrations and Unlabeled Experience
Scaling Language Models: Methods, Analysis & Insights from Training Gopher
ALIGN: Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
Microsoft COCO: Common Objects in Context
Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning
Flamingo: a Visual Language Model for Few-Shot Learning
OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge
VQA: Visual Question Answering
T0: Multitask Prompted Training Enables Zero-Shot Task Generalization
FLAN: Finetuned Language Models Are Zero-Shot Learners
Progressive Neural Networks
Benchmarking End-to-End Behavioral Cloning on Video Games
GPT-3: Language Models are Few-Shot Learners
Scaling Laws for Autoregressive Generative Modeling
Image GPT (iGPT): We find that, just as a large transformer model trained on language can generate coherent text, the same exact model trained on pixel sequences can generate coherent image completions and samples
How AI Training Scales
WebGPT: Browser-assisted question-answering with human feedback
Perceiver: General Perception with Iterative Attention
Open-Ended Learning Leads to Generally Capable Agents
From Motor Control to Team Play in Simulated Humanoid Football
Scaling Laws for Transfer
Scaling Laws for Neural Language Models