Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank
https://sites.google.com/site/offensevalsharedtask/olid
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
XLNet: Generalized Autoregressive Pretraining for Language Understanding
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
https://github.com/UCF-ML-Research/TrojText