-
Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank
-
https://sites.google.com/site/offensevalsharedtask/olid
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
-
XLNet: Generalized Autoregressive Pretraining for Language Understanding
-
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
-
https://github.com/UCF-ML-Research/TrojText
-