-
Attention Is All You Need
-
Reformer: The Efficient Transformer
-
Linformer: Self-Attention with Linear Complexity
-
FAVOR+: Rethinking Attention with Performers
-
Longformer: The Long-Document Transformer
-
https://arxiv.org/pdf/2009.06732.pdf#org=google&page=5
-
https://arxiv.org/pdf/2009.06732.pdf#org=google&page=6
-