Attention Is All You Need
Reformer: The Efficient Transformer
Linformer: Self-Attention with Linear Complexity
FAVOR+: Rethinking Attention with Performers
Longformer: The Long-Document Transformer
https://arxiv.org/pdf/2009.06732.pdf#org=google&page=5
https://arxiv.org/pdf/2009.06732.pdf#org=google&page=6