“‘Transformer Matrix Optimizations’ Tag”,2021-04-17 ():
Bibliography for tag
ai/nn/transformer/attention/linear-algebra, most recent first: 18 annotations & 2 links (parent).
- See Also
- Links
- “LoLCATs: On Low-Rank Linearizing of Large Language Models”, et al 2024
- “SANA: Efficient High-Resolution Image Synthesis With Linear Diffusion Transformers”, et al 2024
- “Conv-Basis: A New Paradigm for Efficient Attention Inference and Gradient Computation in Transformers”, et al 2024
- “RWKV: Reinventing RNNs for the Transformer Era”, et al 2023
- “CosFormer: Rethinking Softmax in Attention”, et al 2022
- “Self-Attention Does Not Need 𝒪(n2) Memory”, 2021
- “Skyformer: Remodel Self-Attention With Gaussian Kernel and Nyström Method”, et al 2021
- “On Learning the Transformer Kernel”, et al 2021
- “A Dot Product Attention Free Transformer”, et al 2021
- “Luna: Linear Unified Nested Attention”, et al 2021
- “Beyond Self-Attention: External Attention Using Two Linear Layers for Visual Tasks (EAMLP)”, et al 2021
- “Sub-Linear Memory: How to Make Performers SLiM”, et al 2020
- “AFT: An Attention Free Transformer”, 2020
- “LambdaNetworks: Modeling Long-Range Interactions without Attention”, 2020
- “Transformers Are RNNs: Fast Autoregressive Transformers With Linear Attention”, et al 2020
- “Linformer: Self-Attention With Linear Complexity”, et al 2020
- “Efficient Attention: Attention With Linear Complexities”, et al 2018
- “Efficient Attention: Attention With Linear Complexities [Blog]”
- Sort By Magic
- Miscellaneous
- Bibliography