Cached Transformers: Improving Transformers with Differentiable Memory Cache
In-context Autoencoder for Context Compression in a Large Language Model
MeMViT: Memory-Augmented Multiscale Vision Transformer for Efficient Long-Term Video Recognition
Perceiver IO: A General Architecture for Structured Inputs & Outputs
TokenLearner: What Can 8 Learned Tokens Do for Images and Videos?
Not All Memories are Created Equal: Learning to Forget by Expiring
Learning to Summarize Long Texts with Memory Compression and Transfer
Set Transformer: A Framework for Attention-based Permutation-Invariant Neural Networks
Phonotactic Reconstruction of Encrypted VoIP Conversations: Hookt on Fon-iks
https://jvns.ca/blog/2013/10/24/day-16-gzip-plus-poetry-equals-awesome/
https://shkspr.mobi/blog/2024/01/compressing-text-into-images/
https%253A%252F%252Farxiv.org%252Fabs%252F2203.08913%2523google.html
https%253A%252F%252Farxiv.org%252Fabs%252F2109.10862%2523openai.html
Perceiver IO: A General Architecture for Structured Inputs & Outputs
https%253A%252F%252Farxiv.org%252Fabs%252F2107.14795%2523deepmind.html
https%253A%252F%252Farxiv.org%252Fabs%252F2103.03206%2523deepmind.html
https%253A%252F%252Farxiv.org%252Fabs%252F1911.05507%2523deepmind.html
Phonotactic Reconstruction of Encrypted VoIP Conversations: Hookt on Fon-iks
Wikipedia Bibliography: