“‘NN Pruning’ Tag”,2019-12-20 (backlinks):
![]()
Bibliography for tag
ai/nn/sparsity/pruning, most recent first: 84 annotations & 9 links (parent).
- See Also
- Links
- “The Super Weight in Large Language Models”, et al 2024
- “What Matters in Transformers? Not All Attention Is Needed”, et al 2024
- “When Parts Are Greater Than Sums: Individual LLM Components Can Outperform Full Models”, et al 2024
- “Pre-Training Small Base LMs With Fewer Tokens”, et al 2024
- “Streamlining Redundant Layers to Compress Large Language Models”, et al 2024
- “The Unreasonable Ineffectiveness of the Deeper Layers”, et al 2024
- “Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression”, et al 2024
- “SliceGPT: Compress Large Language Models by Deleting Rows and Columns”, et al 2024
- “Weight Subcloning: Direct Initialization of Transformers Using Larger Pretrained Ones”, et al 2023
- “To Grok or Not to Grok: Disentangling Generalization and Memorization on Corrupted Algorithmic Datasets”, et al 2023
- “Sheared LLaMA: Accelerating Language Model Pre-Training via Structured Pruning”, et al 2023
- “One Wide Feedforward Is All You Need”, et al 2023
- “A Comparative Study between Full-Parameter and LoRA-Based Fine-Tuning on Chinese Instruction Data for Instruction Following Large Language Model”, et al 2023
- “Fast As CHITA: Neural Network Pruning With Combinatorial Optimization”, et al 2023
- “Self-Compressing Neural Networks”, 2023
- “Pruning Compact ConvNets for Efficient Inference”, et al 2023
- “Rethinking the Role of Scale for In-Context Learning: An Interpretability-Based Case Study at 66 Billion Scale”, et al 2022
- “Lottery Tickets on a Data Diet: Finding Initializations With Sparse Trainable Networks”, et al 2022
- “Heavy-Tailed Neuronal Connectivity Arises from Hebbian Self–organization”, et al 2022
- “PPCD-GAN: Progressive Pruning and Class-Aware Distillation for Large-Scale Conditional GANs Compression”, et al 2022
- “The Combinatorial Brain Surgeon: Pruning Weights That Cancel One Another in Neural Networks”, et al 2022
- “Data-Efficient Structured Pruning via Submodular Optimization”, et al 2022
- “Sparsity Winning Twice: Better Robust Generalization from More Efficient Training”, et al 2022
- “Fortuitous Forgetting in Connectionist Networks”, et al 2022
- “How Many Degrees of Freedom Do We Need to Train Deep Networks: a Loss Landscape Perspective”, et al 2021
- “Prune Once for All: Sparse Pre-Trained Language Models”, et al 2021
- “DSEE: Dually Sparsity-Embedded Efficient Tuning of Pre-Trained Language Models”, et al 2021
- “HALP: Hardware-Aware Latency Pruning”, et al 2021
- “On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis”, et al 2021
- “Block Pruning For Faster Transformers”, et al 2021
- “Scaling Laws for Deep Learning”, 2021
- “A Winning Hand: Compressing Deep Networks Can Improve Out-Of-Distribution Robustness”, et al 2021
- “Chasing Sparsity in Vision Transformers: An End-To-End Exploration”, et al 2021
- “On Lottery Tickets and Minimal Task Representations in Deep Reinforcement Learning”, et al 2021
- “Sifting out the Features by Pruning: Are Convolutional Networks the Winning Lottery Ticket of Fully Connected Ones?”, 2021
- “Learning N:M Fine-Grained Structured Sparse Neural Networks From Scratch”, et al 2021
- “Postnatal Connectomic Development of Inhibition in Mouse Barrel Cortex”, et al 2021
- “ES-ENAS: Blackbox Optimization over Hybrid Spaces via Combinatorial and Continuous Evolution”, et al 2021
- “Progressively Stacking 2.0: A Multi-Stage Layerwise Training Method for BERT Training Speedup”, et al 2020
- “A Primer in BERTology: What We Know about How BERT Works”, et al 2020
- “Bort: Optimal Subarchitecture Extraction For BERT”, 2020
- “Pruning Neural Networks at Initialization: Why Are We Missing the Mark?”, et al 2020
- “Logarithmic Pruning Is All You Need”, et al 2020
- “On the Predictability of Pruning Across Scales”, et al 2020
- “Progressive Skeletonization: Trimming More Fat from a Network at Initialization”, et al 2020
- “Pruning Neural Networks without Any Data by Iteratively Conserving Synaptic Flow”, et al 2020
- “Movement Pruning: Adaptive Sparsity by Fine-Tuning”, et al 2020
- “Bayesian Bits: Unifying Quantization and Pruning”, et al 2020
- “Lite Transformer With Long-Short Range Attention”, et al 2020
- “On the Effect of Dropping Layers of Pre-Trained Transformer Models”, et al 2020
- “Train-By-Reconnect: Decoupling Locations of Weights from Their Values (LaPerm)”, 2020
- “Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers”, et al 2020
- “What’s Hidden in a Randomly Weighted Neural Network?”, et al 2019
- “Sparse Networks from Scratch: Faster Training without Losing Performance”, 2019
- “Playing the Lottery With Rewards and Multiple Languages: Lottery Tickets in RL and NLP”, et al 2019
- “SpArSe: Sparse Architecture Search for CNNs on Resource-Constrained Microcontrollers”, et al 2019
- “Are 16 Heads Really Better Than One?”, et al 2019
- “Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned”, et al 2019
- “Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask”, et al 2019
- “Stabilizing the Lottery Ticket Hypothesis”, et al 2019
- “The State of Sparsity in Deep Neural Networks”, et al 2019
- “Differential Contribution of Cortical Thickness, Surface Area, and Gyrification to Fluid and Crystallized Intelligence”, et al 2019
- “Efficient Training of BERT by Progressively Stacking”, et al 2019
- “A Closer Look at Structured Pruning for Neural Network Compression”, et al 2018
- “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks”, 2018
- “Efficient Neural Audio Synthesis”, et al 2018
- “Recovering from Random Pruning: On the Plasticity of Deep Convolutional Neural Networks”, et al 2018
- “Learning to Prune Filters in Convolutional Neural Networks”, et al 2018
- “Faster Gaze Prediction With Dense Networks and Fisher Pruning”, et al 2018
- “Automated Pruning for Deep Neural Network Compression”, et al 2017
- “Training Simplification and Model Simplification for Deep Learning: A Minimal Effort Back Propagation Method”, et al 2017
- “NeST: A Neural Network Synthesis Tool Based on a Grow-And-Prune Paradigm”, et al 2017
- “To Prune, or Not to Prune: Exploring the Efficacy of Pruning for Model Compression”, 2017
- “Bayesian Sparsification of Recurrent Neural Networks”, et al 2017
- “Structured Bayesian Pruning via Log-Normal Multiplicative Noise”, et al 2017
- “Exploring Sparsity in Recurrent Neural Networks”, et al 2017
- “Variational Dropout Sparsifies Deep Neural Networks”, et al 2017
- “Iterative Magnitude Pruning: Learning Both Weights and Connections for Efficient Neural Networks”, et al 2015
- “Flat Minima”, 1997
- “Optimal Brain Surgeon and General Network Pruning”, et al 1993
- “Fault Tolerance of Pruned Multilayer Networks”, 1991
- “Using Relevance to Reduce Network Size Automatically”, 1989
- “Optimal Brain Damage”, et al 1989
- “Trading Off Compute in Training and Inference § Pruning”
- Wikipedia
- Miscellaneous
- Bibliography