BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
CLIP: Connecting Text and Images: We’re introducing a neural network called CLIP which efficiently learns visual concepts from natural language supervision. CLIP can be applied to any visual classification benchmark by simply providing the names of the visual categories to be recognized, similar to the ‘zero-shot’ capabilities of GPT-2 and GPT-3
IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
R2D2: Recurrent Experience Replay in Distributed Reinforcement Learning
Never Give Up: Learning Directed Exploration Strategies
RND: Large-Scale Study of Curiosity-Driven Learning