Bibliography (17):

  1. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

  2. DALL·E 1: Creating Images from Text: We’ve trained a neural network called DALL·E that creates images from text captions for a wide range of concepts expressible in natural language

  3. GPT-3: Language Models are Few-Shot Learners

  4. https://www.alignmentforum.org/posts/Haawpd5rZrzkzvYRC/an-162-foundation-models-a-paradigm-shift-within-ai

  5. https://arxiv.org/pdf/2108.07258.pdf#page=26

  6. https://arxiv.org/pdf/2108.07258.pdf#page=34

  7. https://arxiv.org/pdf/2108.07258.pdf#page=42

  8. https://arxiv.org/pdf/2108.07258.pdf#page=54

  9. https://arxiv.org/pdf/2108.07258.pdf#page=85

  10. The Power of Scale for Parameter-Efficient Prompt Tuning

  11. https://mailchi.mp/459b1e4f860d/an-152how-weve-overestimated-few-shot-learning-capabilities

  12. https://thegradient.pub/prompting/

  13. https://mailchi.mp/aa6782968981/an-155a-minecraft-benchmark-for-algorithms-that-learn-without-reward-functions

  14. https://arxiv.org/pdf/2108.07258.pdf#page=92

  15. https://arxiv.org/pdf/2108.07258.pdf#page=114

  16. https://arxiv.org/pdf/2108.07258.pdf#page=118