Bibliography (6):

  1. Attention Is All You Need

  2. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

  3. Unsupervised Cross-lingual Representation Learning at Scale

  4. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding