Bibliography (30):

  1. Chinese AI lab challenges Google, OpenAI with a model of 1.75 trillion parameters

  2. GPT-3: Language Models are Few-Shot Learners

  3. DALL·E 1: Creating Images from Text: We’ve trained a neural network called DALL·E that creates images from text captions for a wide range of concepts expressible in natural language

  4. Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity

  5. https://scholar.google.com/citations?user=n1zDCkQAAAAJ&view_op=list_works&alert_preview_top_rm=2&sortby=pubdate

  6. https://scholar.google.com/citations?user=GQWTo4MAAAAJ

  7. China’s AI Advances Help Its Tech Industry, and State Security

  8. https://scholar.google.com/citations?user=y5zkBeMAAAAJ

  9. China’s GPT-3? BAAI Introduces Superscale Intelligence Model ‘Wu Dao 1.0’: The Beijing Academy of Artificial Intelligence (BAAI) releases Wu Dao 1.0, China’s first large-scale pretraining model.

  10. ChinAI #141: The PanGu Origin Story: Notes from an informative Zhihu Thread on PanGu

  11. Turing-NLG: A 17-billion-parameter language model by Microsoft

  12. CPM: A Large-scale Generative Chinese Pre-trained Language Model

  13. [Ali released PLUG: 27 billion parameters, the largest pre-trained language model in the Chinese community]

  14. DeepSpeed: Extreme-scale model training for everyone

  15. ERNIE 3.0: Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation

  16. Yuan 1.0: Large-Scale Pre-trained Language Model in Zero-Shot and Few-Shot Learning

  17. ERNIE 3.0 Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation

  18. Microsoft announces new supercomputer, lays out vision for future AI work

  19. Exploring Sparse Expert Models and Beyond