Attention Is All You Need
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Unsupervised Cross-lingual Representation Learning at Scale
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding