Bibliography:

  1. ‘neural net’ tag

  2. ‘multi-scale Transformers’ tag

  3. ‘autoencoder NN’ tag

  4. ‘AI scaling’ tag

  5. ‘AlphaGo’ tag

  6. InvertOrNot.com Proposal

  7. Convolutional Differentiable Logic Gate Networks

  8. MaskBit: Embedding-free Image Generation via Bit Tokens

  9. Quantum Convolutional Neural Networks are (Effectively) Classically Simulable

  10. Three-Dimension Animation Character Design Based on Probability Genetic Algorithm

  11. Investigating learning-independent abstract reasoning in artificial neural networks

  12. Grokfast: Accelerated Grokking by Amplifying Slow Gradients

  13. A Rotation and a Translation Suffice: Fooling CNNs with Simple Transformations

  14. Neural Networks Learn Statistics of Increasing Complexity

  15. Machine learning reveals the control mechanics of an insect wing hinge

  16. Supplementary Materials for Grounded language acquisition through the eyes and ears of a single child

  17. Grounded language acquisition through the eyes and ears of a single child

  18. Machine Learning as a Tool for Hypothesis Generation

  19. Multi visual feature fusion based fog visibility estimation for expressway surveillance using deep learning network

  20. Auditing the inference processes of medical-image classifiers by leveraging generative AI and the expertise of physicians

  21. Development of Deep Ensembles to Screen for Autism and Symptom Severity Using Retinal Photographs

  22. May the Noise be with you: Adversarial Training without Adversarial Examples

  23. Are Vision Transformers More Data Hungry Than Newborn Visual Systems?

  24. UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition

  25. The possibility of making $138,000 from shredded banknote pieces using computer vision

  26. ConvNets Match Vision Transformers at Scale

  27. Interpret Vision Transformers as ConvNets with Dynamic Convolutions

  28. Artificial intelligence-supported screen reading versus standard double reading in the Mammography Screening with Artificial Intelligence trial (MASAI): a clinical safety analysis of a randomised, controlled, non-inferiority, single-blinded, screening accuracy study

  29. Hand-drawn anime line drawing colorization of faces with texture details

  30. High-Quality Synthetic Character Image Extraction via Distortion Recognition

  31. Loss of Plasticity in Deep Continual Learning (Continual Backpropagation)

  32. Neural networks trained with SGD learn distributions of increasing complexity

  33. Rosetta Neurons: Mining the Common Units in a Model Zoo

  34. Improving neural network representations using human similarity judgments

  35. U-Net CNN in APL: Exploring Zero-Framework, Zero-Library Machine Learning

  36. VanillaNet: the Power of Minimalism in Deep Learning

  37. Multi-Label Classification in Anime Illustrations Based on Hierarchical Attribute Relationships

  38. ImageNet-Hard: The Hardest Images Remaining from a Study of the Power of Zoom and Spatial Biases in Image Classification

  39. Hierarchical Multi-Label Attribute Classification With Graph Convolutional Networks on Anime Illustration

  40. Loss Landscapes are All You Need: Neural Network Generalization Can Be Explained Without the Implicit Bias of Gradient Descent

  41. Adding Conditional Control to Text-to-Image Diffusion Models

  42. Pruning Compact ConvNets for Efficient Inference

  43. Does progress on ImageNet transfer to real-world datasets?

  44. EarSpy: Spying Caller Speech and Identity through Tiny Vibrations of Smartphone Ear Speakers

  45. Pretraining Without Attention

  46. What do Vision Transformers Learn? A Visual Exploration

  47. A 64-core mixed-signal in-memory compute chip based on phase-change memory for deep neural network inference

  48. Simulated automated facial recognition systems as decision-aids in forensic face matching tasks

  49. Interpreting Neural Networks through the Polytope Lens

  50. Predicting sex, age, general cognition and mental health with machine learning on brain structural connectomes

  51. The Power of Ensembles for Active Learning in Image Classification

  52. GCN-Based Multi-Modal Multi-Label Attribute Classification in Anime Illustration Using Domain-Specific Semantic Features

  53. The Unreasonable Effectiveness of Fully-Connected Layers for Low-Data Regimes

  54. Understanding the Covariance Structure of Convolutional Filters

  55. VICRegL: Self-Supervised Learning of Local Visual Features

  56. Omnigrok: Grokking Beyond Algorithmic Data

  57. g.pt: Learning to Learn with Generative Models of Neural Network Checkpoints

  58. FastSiam: Resource-Efficient Self-supervised Learning on a Single GPU

  59. Evaluation of Transfer Learning Methods for Detecting Alzheimer’s Disease with Brain MRI

  60. Reassessing hierarchical correspondences between brain and deep networks through direct interface

  61. Timesweeper: Accurately Identifying Selective Sweeps Using Population Genomic Time Series

  62. RHO-LOSS: Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt

  63. BigVGAN: A Universal Neural Vocoder with Large-Scale Training

  64. Studying Growth With Neural Cellular Automata

  65. Continual Pre-Training Mitigates Forgetting in Language and Vision

  66. Scaling Up Your Kernels to 31×31: Revisiting Large Kernel Design in CNNs (RepLKNet)

  67. Democratizing Contrastive Language-Image Pre-training: A CLIP Benchmark of Data, Model, and Supervision

  68. Variational Autoencoders Without the Variation

  69. On the Effectiveness of Dataset Watermarking in Adversarial Settings

  70. General Cyclical Training of Neural Networks

  71. Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet

  72. Variational Neural Cellular Automata

  73. ConvMixer: Patches Are All You Need?

  74. HyperTransformer: Model Generation for Supervised and Semi-Supervised Few-Shot Learning

  75. ConvNeXt: A ConvNet for the 2020s

  76. An Empirical Investigation of the Role of Pre-training in Lifelong Learning

  77. Noether Networks: Meta-Learning Useful Conserved Quantities

  78. AugMax: Adversarial Composition of Random Augmentations for Robust Training

  79. The Efficiency Misnomer

  80. Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators

  81. Evaluating Loss Functions for Illustration Super-Resolution Neural Networks

  82. TWIST: Self-Supervised Learning by Estimating Twin Class Distributions

  83. Deep learning models of cognitive processes constrained by human brain connectomes

  84. Decoupled Contrastive Learning

  85. Supervision Exists Everywhere: A Data Efficient Contrastive Language-Image Pre-training Paradigm (DeCLIP)

  86. Mining for strong gravitational lenses with self-supervised learning

  87. THINGSvision: A Python Toolbox for Streamlining the Extraction of Activations From Deep Neural Networks

  88. A Battle of Network Structures: An Empirical Study of CNN, Transformer, and MLP

  89. Predicting phenotypes from genetic, environment, management, and historical data using CNNs

  90. Do Vision Transformers See Like Convolutional Neural Networks?

  91. Dataset Distillation with Infinitely Wide Convolutional Networks

  92. Neuroprosthesis for Decoding Speech in a Paralyzed Person with Anarthria

  93. Graph Jigsaw Learning for Cartoon Face Recognition

  94. Prediction Depth: Deep Learning Through the Lens of Example Difficulty

  95. Revisiting the Calibration of Modern Neural Networks

  96. Partial success in closing the gap between human and machine vision

  97. CoAtNet: Marrying Convolution and Attention for All Data Sizes

  98. Effect of Pre-Training Scale on Intra/Inter-Domain Full and Few-Shot Transfer Learning for Natural and Medical X-Ray Chest Images

  99. Embracing New Techniques in Deep Learning for Estimating Image Memorability

  100. Predicting sex from retinal fundus photographs using automated deep learning

  101. Rethinking and Improving the Robustness of Image Style Transfer

  102. Rip van Winkle’s Razor, a Simple New Estimate for Adaptive Data Analysis

  103. The surprising impact of mask-head architecture on novel class segmentation

  104. Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks

  105. ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases

  106. Learning from videos to understand the world

  107. Fast and Accurate Model Scaling

  108. Transfer of Fully Convolutional Policy-Value Networks Between Games and Game Variants

  109. Momentum Residual Neural Networks

  110. Hiding Data Hiding

  111. Explaining Neural Scaling Laws

  112. NFNet: High-Performance Large-Scale Image Recognition Without Normalization

  113. Brain2Pix: Fully convolutional naturalistic video reconstruction from brain activity

  114. Words as a window: Using word embeddings to explore the learned representations of Convolutional Neural Networks

  115. E(3)-Equivariant Graph Neural Networks for Data-Efficient and Accurate Interatomic Potentials

  116. Meta Pseudo Labels

  117. Is MLP-Mixer a CNN in Disguise? As Part of This Blog Post, We Look at the MLP Mixer Architecture in Detail and Also Understand Why It Is Not Considered Convolution Free.

  118. Converting tabular data into images for deep learning with convolutional neural networks

  119. Taming Transformers for High-Resolution Image Synthesis

  120. Ensemble Learning of Convolutional Neural Network, Support Vector Machine, and Best Linear Unbiased Predictor for Brain Age Prediction: ARAMIS Contribution to the Predictive Analytics Competition 2019 Challenge

  121. Scaling down Deep Learning

  122. Very Deep VAEs Generalize Autoregressive Models and Can Outperform Them on Images

  123. Understanding RL Vision: With diverse environments, we can analyze, diagnose and edit deep reinforcement learning models using attribution

  124. Fourier Neural Operator for Parametric Partial Differential Equations

  125. Deep learning-based classification of the polar emotions of ‘moe’-style cartoon pictures

  126. Sharpness-Aware Minimization (SAM) for Efficiently Improving Generalization

  127. Demonstrating that dataset domains are largely linearly separable in the feature space of common CNNs

  128. Optimal Peanut Butter and Banana Sandwiches

  129. Accuracy and Performance Comparison of Video Action Recognition Approaches

  130. A digital biomarker of diabetes from smartphone-based vascular signals

  131. SSD vs. YOLO for Detection of Outdoor Urban Advertising Panels under Multiple Variabilities

  132. On Robustness and Transferability of Convolutional Neural Networks

  133. NVAE: A Deep Hierarchical Variational Autoencoder

  134. CoCoNuT: Combining Context-Aware Neural Translation Models using Ensemble for Program Repair

  135. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization

  136. SimCLRv2: Big Self-Supervised Models are Strong Semi-Supervised Learners

  137. FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining

  138. Danny Hernandez on forecasting and the drivers of AI progress

  139. Measuring the Algorithmic Efficiency of Neural Networks

  140. AI and Efficiency: We’re releasing an analysis showing that since 2012 the amount of compute needed to train a neural net to the same performance on ImageNet classification has been decreasing by a factor of 2 every 16 months

  141. Reinforcement Learning with Augmented Data

  142. YOLOv4: Optimal Speed and Accuracy of Object Detection

  143. Scaling Laws from the Data Manifold Dimension

  144. Shortcut Learning in Deep Neural Networks

  145. Evolving Normalization-Activation Layers

  146. Conditional Convolutions for Instance Segmentation

  147. Train-by-Reconnect: Decoupling Locations of Weights from their Values (LaPerm)

  148. Rethinking Parameter Counting in Deep Models: Effective Dimensionality Revisited

  149. Do We Need Zero Training Loss After Achieving Zero Training Error?

  150. Bayesian Deep Learning and a Probabilistic Perspective of Generalization

  151. A Simple Framework for Contrastive Learning of Visual Representations

  152. Growing Neural Cellular Automata: Differentiable Model of Morphogenesis

  153. Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving

  154. First-in-human evaluation of a hand-held automated venipuncture device for rapid venous blood draws

  155. ImageNet-A: Natural Adversarial Examples

  156. Deep-Eyes: Fully Automatic Anime Character Colorization with Painting of Details on Empty Pupils

  157. CARN: Convolutional Anchored Regression Network for Fast and Accurate Single Image Super-Resolution

  158. 3fe19144f573eaba2ca1b17d9880bd983a0d0d9a.pdf

  159. The Importance of Deconstruction

  160. Big Transfer (BiT): General Visual Representation Learning

  161. Linear Mode Connectivity and the Lottery Ticket Hypothesis

  162. Dynamic Convolution: Attention over Convolution Kernels

  163. Deep Double Descent: We show that the double descent phenomenon occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or training time

  164. Fantastic Generalization Measures and Where to Find Them

  165. Anonymous market product classification based on deep learning

  166. The Origins and Prevalence of Texture Bias in Convolutional Neural Networks

  167. How Machine Learning Can Help Unlock the World of Ancient Japan

  168. SimpleShot: Revisiting Nearest-Neighbor Classification for Few-Shot Learning

  169. Self-training with Noisy Student improves ImageNet classification

  170. Taxonomy of Real Faults in Deep Learning Systems

  171. On the Measure of Intelligence

  172. DD-PPO: Learning Near-Perfect PointGoal Navigators from 2.5 Billion Frames

  173. Accelerating Deep Learning by Focusing on the Biggest Losers

  174. ANIL: Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML

  175. ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models

  176. CAR: Learned Image Downscaling for Upscaling using Content Adaptive Resampler

  177. Finding the Needle in the Haystack with Convolutions: on the benefits of architectural bias

  178. Intriguing properties of adversarial training at scale

  179. Adversarial Robustness as a Prior for Learned Representations

  180. Human-level performance in 3D multiplayer games with population-based reinforcement learning

  181. ImageNet-Sketch: Learning Robust Global Representations by Penalizing Local Predictive Power

  182. Cold Case: The Lost MNIST Digits

  183. Improved object recognition using neural networks trained to mimic the brain’s statistical properties

  184. Neural System Identification with Neural Information Flow

  185. Percival: Making In-Browser Perceptual Ad Blocking Practical With Deep Learning

  186. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features

  187. Adversarial Examples Are Not Bugs, They Are Features

  188. Searching for MobileNetV3

  189. Billion-scale semi-supervised learning for image classification

  190. A Recipe for Training Neural Networks

  191. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection

  192. COCO-GAN: Generation by Parts via Conditional Coordinating

  193. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

  194. Semantic Image Synthesis with Spatially-Adaptive Normalization

  195. The Bitter Lesson

  196. Learning To Follow Directions in Street View

  197. SuperTML: Two-Dimensional Word Embedding for the Precognition on Structured Tabular Data

  198. Real-time Continuous Transcription with Live Transcribe

  199. Do We Train on Test Data? Purging CIFAR of Near-Duplicates

  200. Pay Less Attention with Lightweight and Dynamic Convolutions

  201. Association Between Surgical Skin Markings in Dermoscopic Images and Diagnostic Performance of a Deep Learning Convolutional Neural Network for Melanoma Recognition

  202. Detecting advertising on building façades with computer vision

  203. On Lazy Training in Differentiable Programming

  204. Quantifying Generalization in Reinforcement Learning

  205. ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware

  206. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness

  207. Evolving Space-Time Neural Architectures for Videos

  208. ADNet: A Deep Network for Detecting Adverts

  209. AdVersarial: Perceptual Ad Blocking meets Adversarial Machine Learning

  210. FloWaveNet: A Generative Flow for Raw Audio

  211. StreetNet: Preference Learning with Convolutional Neural Network on Urban Crime Perception

  212. Towards Understanding Learning Representations: To What Extent Do Different Neural Networks Learn the Same Representation

  213. Understanding and correcting pathologies in the training of learned optimizers

  214. Graph Convolutional Reinforcement Learning

  215. Cellular automata as convolutional neural networks

  216. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization

  217. Human-Like Playtesting with Deep Learning

  218. CurriculumNet: Weakly Supervised Learning from Large-Scale Web Images

  219. MnasNet: Platform-Aware Neural Architecture Search for Mobile

  220. LEO: Meta-Learning with Latent Embedding Optimization

  221. Glow: Generative Flow with Invertible 1×1 Convolutions

  222. The Goldilocks zone: Towards better understanding of neural network loss landscapes

  223. Benchmarking Neural Network Robustness to Common Corruptions and Surface Variations

  224. Confounding variables can degrade generalization performance of radiological deep learning models

  225. Faster SGD training by minibatch persistency

  226. Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks

  227. Resource-Efficient Neural Architect

  228. More Than a Feeling: Learning to Grasp and Regrasp using Vision and Touch

  229. Deep learning generalizes because the parameter-function map is biased towards simple functions

  230. Bidirectional Learning for Robust Neural Networks

  231. BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning

  232. Self-distillation: Born Again Neural Networks

  233. Tile2Vec: Unsupervised representation learning for spatially distributed data

  234. Exploring the Limits of Weakly Supervised Pretraining

  235. YOLOv3: An Incremental Improvement

  236. Reptile: On First-Order Meta-Learning Algorithms

  237. Essentially No Barriers in Neural Network Energy Landscape

  238. Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs

  239. Guess, check and fix: a phenomenology of improvisation in ‘neural’ painting

  240. Sim-to-Real Optimization of Complex Real World Mobile Network with Imperfect Information via Deep Reinforcement Learning from Self-play

  241. Evolved Policy Gradients

  242. Large-scale, high-resolution comparison of the core visual object recognition behavior of humans, monkeys, and state-of-the-art deep artificial neural networks

  243. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures

  244. Active, Continual Fine Tuning of Convolutional Neural Networks for Reducing Annotation Efforts

  245. ArcFace: Additive Angular Margin Loss for Deep Face Recognition

  246. Man against Machine: Diagnostic Performance of a Deep Learning Convolutional Neural Network for Dermoscopic Melanoma Recognition in Comparison to 58 Dermatologists

  247. DeepGS: Predicting phenotypes from genotypes using Deep Learning

  248. Deep image reconstruction from human brain activity

  249. Visualizing the Loss Landscape of Neural Nets

  250. SPP-Net: Deep Absolute Pose Regression with Synthetic Views

  251. China’s AI Advances Help Its Tech Industry, and State Security

  252. Measuring the tendency of CNNs to Learn Surface Statistical Regularities

  253. 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks

  254. BlockDrop: Dynamic Inference Paths in Residual Networks

  255. Knowledge Concentration: Learning 100K Object Classifiers in a Single CNN

  256. The signature of robot action success in EEG signals of a human observer: Decoding and visualization using deep convolutional neural networks

  257. 11K Hands: Gender recognition and biometric identification using a large dataset of hand images

  258. Learning to Play Chess with Minimal Lookahead and Deep Value Neural Networks

  259. Learning to Generalize: Meta-Learning for Domain Generalization

  260. High-Precision Automated Reconstruction of Neurons with Flood-filling Networks

  261. Efficient K-shot Learning with Regularized Deep Networks

  262. NIMA: Neural Image Assessment

  263. Squeeze-and-Excitation Networks

  264. What does a convolutional neural network recognize in the moon?

  265. SMASH: One-Shot Model Architecture Search through HyperNetworks

  266. BitNet: Bit-Regularized Deep Neural Networks

  267. A deep architecture for unified esthetic prediction

  268. Learning with Rethinking: Recurrently Improving Convolutional Neural Networks through Feedback

  269. WebVision Database: Visual Learning and Understanding from Web Data

  270. Focal Loss for Dense Object Detection

  271. Active Learning for Convolutional Neural Networks: A Core-Set Approach

  272. Learning to Infer Graphics Programs from Hand-Drawn Images

  273. A Downsampled Variant of ImageNet as an Alternative to the CIFAR datasets

  274. Learning Transferable Architectures for Scalable Image Recognition

  275. Efficient Architecture Search by Network Transformation

  276. A Simple Neural Attentive Meta-Learner

  277. Towards Deep Learning Models Resistant to Adversarial Attacks

  278. Gradient Diversity: a Key Ingredient for Scalable Distributed Learning

  279. Device Placement Optimization with Reinforcement Learning

  280. Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour

  281. Submanifold Sparse Convolutional Networks

  282. A simple neural network module for relational reasoning

  283. Deep Learning is Robust to Massive Label Noise

  284. What Makes a Good Image? Airbnb Demand Analytics Leveraging Interpretable Image Features

  285. Picasso: A Modular Framework for Visualizing the Learning Process of Neural Network Image Classifiers

  286. BAM! The Behance Artistic Media Dataset for Recognition Beyond Photography

  287. Adversarial Neural Machine Translation

  288. Multi-Scale Dense Networks for Resource Efficient Image Classification

  289. Scaling the Scattering Transform: Deep Hybrid Networks

  290. Mask R-CNN

  291. Using Human Brain Activity to Guide Machine Learning

  292. Learned Optimizers that Scale and Generalize

  293. Prediction and Control with Temporal Segment Models

  294. Parallel Multiscale Autoregressive Density Estimation

  295. Convolution Aware Initialization

  296. Gender-From-Iris or Gender-From-Mascara?

  297. BrainNetCNN: Convolutional neural networks for brain networks; towards predicting neurodevelopment

  298. Universal representations: The missing link between faces, text, planktons, and cat breeds

  299. PixelCNN++: Improving the PixelCNN with Discretized Logistic Mixture Likelihood and Other Modifications

  300. YOLO9000: Better, Faster, Stronger

  301. Language Modeling with Gated Convolutional Networks

  302. Learning from Simulated and Unsupervised Images through Adversarial Training

  303. LipNet: End-to-End Sentence-level Lipreading

  304. Feature Pyramid Networks for Object Detection

  305. Self-critical Sequence Training for Image Captioning

  306. ResNeXt: Aggregated Residual Transformations for Deep Neural Networks

  307. Responses to Critiques on Machine Learning of Criminality Perceptions (Addendum of arXiv:1611.04135)

  308. Understanding deep learning requires rethinking generalization

  309. Designing Neural Network Architectures using Reinforcement Learning

  310. VPN: Video Pixel Networks

  311. HyperNetworks

  312. Neural Photo Editing with Introspective Adversarial Networks

  313. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima

  314. WaveNet: A Generative Model for Raw Audio

  315. Direct Feedback Alignment Provides Learning in Deep Neural Networks

  316. Deep Learning Human Mind for Automated Visual Classification

  317. Temporal Convolutional Networks: A Unified Approach to Action Segmentation

  318. DenseNet: Densely Connected Convolutional Networks

  319. Clockwork Convnets for Video Semantic Segmentation

  320. Deep Learning the City: Quantifying Urban Perception At A Global Scale

  321. Convolutional Neural Fabrics

  322. Deep neural networks are robust to weight binarization and other non-linear distortions

  323. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution (ASPP), and Fully Connected CRFs

  324. FractalNet: Ultra-Deep Neural Networks without Residuals

  325. Wide Residual Networks

  326. Residual Networks Behave Like Ensembles of Relatively Shallow Networks

  327. Neural Autoregressive Distribution Estimation

  328. ViZDoom: A Doom-based AI Research Platform for Visual Reinforcement Learning

  329. OHEM: Training Region-based Object Detectors with Online Hard Example Mining

  330. Deep Networks with Stochastic Depth

  331. Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing

  332. Do Deep Convolutional Nets Really Need to be Deep and Convolutional?

  333. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks

  334. Learning Hand-Eye Coordination for Robotic Grasping with Deep Learning and Large-Scale Data Collection

  335. Network Morphism

  336. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning

  337. PlaNet—Photo Geolocation with Convolutional Neural Networks

  338. Value Iteration Networks

  339. PixelRNN: Pixel Recurrent Neural Networks

  340. Image Synthesis from Yahoo’s open_nsfw

  341. Deep Residual Learning for Image Recognition

  342. Microsoft researchers win ImageNet computer vision challenge

  343. Adding Gradient Noise Improves Learning for Very Deep Networks

  344. The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition

  345. Learning Visual Features from Large Weakly Supervised Data

  346. Illustration2Vec: a semantic vector representation of illustrations

  347. BinaryConnect: Training Deep Neural Networks with binary weights during propagations

  348. Predicting and Understanding Urban Perception with Convolutional Neural Networks

  349. A Neural Attention Model for Abstractive Sentence Summarization

  350. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop

  351. You Only Look Once: Unified, Real-Time Object Detection

  352. Clothing-1M: Learning from Massive Noisy Labeled Data for Image Classification

  353. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning

  354. STN: Spatial Transformer Networks

  355. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

  356. Cyclical Learning Rates for Training Neural Networks

  357. Deep Learning

  358. Fast R-CNN

  359. End-to-End Training of Deep Visuomotor Policies

  360. FaceNet: A Unified Embedding for Face Recognition and Clustering

  361. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

  362. DeepID3: Face Recognition with Very Deep Neural Networks

  363. Explaining and Harnessing Adversarial Examples

  364. Understanding image representations by measuring their equivariance and equivalence

  365. Going Deeper with Convolutions

  366. Very Deep Convolutional Networks for Large-Scale Image Recognition

  367. ImageNet Large Scale Visual Recognition Challenge

  368. Deep Learning Face Representation by Joint Identification-Verification

  369. One weird trick for parallelizing convolutional neural networks

  370. Network In Network

  371. R-CNN: Rich feature hierarchies for accurate object detection and semantic segmentation

  372. Maxout Networks

  373. ImageNet Classification with Deep Convolutional Neural Networks

  374. Multi-column deep neural network for traffic sign classification

  375. Multi-column Deep Neural Networks for Image Classification

  376. Building high-level features using large scale unsupervised learning

  377. DanNet: Flexible, High Performance Convolutional Neural Networks for Image Classification

  378. Hypernetworks [Blog]

  379. Deconvolution and Checkerboard Artifacts

  380. Hierarchical Object Detection With Deep Reinforcement Learning

  381. f106bf397cea4b3c184a40c91893ee695f7646df.html

  382. Creating a 17 KB Style Transfer Model With Layer Pruning and Quantization

  383. Now Anyone Can Train Imagenet in 18 Minutes

  384. Cats, Rats, A.I., Oh My!

  385. 2023-hsu-appendix-acompleteunetnnimplementationinapl.png

  386. 2022-mindermann-figure1-18xspeedupfromactivelearningofclothing1mdataset.jpg

  387. 2022-smith-table1-cyclicalweightdecaycifarimagenet.png

  388. 2020-07-24-gwern-meme-moneyprinter-bitterlesson-gpt3.png

  389. 2019-boazbarak-deepdoubledescent-expandedversion-degree1000spline-goodoverfitting.png

  390. 2019-humbatova-figure1-taxonomyofrealfaultsindeeplearningsystems.png

  391. 2018-mahajan-figure2-imagenetcub2011transferlearningfrominstagramhashtagsscalingcurves.jpg

  392. 2017-rawat.pdf

  393. 2015-joulin-figure2-flickrpascalvoc2007precisionscalingwithflickr100mnscaling.jpg

  394. https://aclanthology.org/D13-1176.pdf

  395. 8ecd1e30c66400dbf395e3f5b7852c8bb70a02d2.pdf

  396. https://animatedai.github.io/

  397. https://frankzliu.com/blog/vision-transformers-are-overrated

  398. ce5cab2244b861ebaefca3f9b0a377b832d07921.html

  399. https://github.com/lukas-blecher/LaTeX-OCR

  400. https://github.com/ultralytics/yolov5/issues/6998#issue-1170533269

  401. https://heritagesciencejournal.springeropen.com/articles/10.1186/s40494-023-01094-0

  402. a61009d7a51e4cad361ec5fa9ea1b9298e69b6a0.html

  403. https://invertornot.com/

  404. https://karpathy.github.io/2022/03/14/lecun1989/

  405. https://myrtle.ai/learn/how-to-train-your-resnet/

  406. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4588941

  407. https://schollz.com/tinker/watercolor/

  408. fb6541535fe7e5e0059b867451c7e3e8ed692b17.html

  409. https://wandb.ai/wandb_fc/articles/reports/Image-to-LaTeX--Vmlldzo1NDQ0MTAx

  410. https://www.jerpint.io/blog/snowflake-neural-ca/

  411. 7da551ce8c6fb8cdab7d63eec915ec35b6cd26c5.html

  412. https://www.nature.com/articles/s41467-023-41693-w

  413. https://www.nature.com/articles/s42003-023-05098-1

  414. https://www.quantamagazine.org/sparse-neural-networks-point-physicists-to-useful-data-20230608/

  415. 6f5ad29694d194976fd23467d00a4bbabcbca622.html

  416. https://www.reddit.com/r/MachineLearning/comments/jthxui/p_chasing_intruding_cats_from_your_home_with/

  417. https://www.reddit.com/r/mlscaling/comments/1ggr0j4/neural_network_recognizer_for_handwritten_zip/

  418. https://www.wired.com/story/beauty-is-in-the-eye-of-the-beholder-but-memorability-may-be-universal/

  419. https://x.com/chriswolfvision/status/1313059518574718977

  420. MaskBit: Embedding-free Image Generation via Bit Tokens

  421. https%253A%252F%252Farxiv.org%252Fabs%252F2409.16211%2523bytedance.html

  422. Grokfast: Accelerated Grokking by Amplifying Slow Gradients

  423. https%253A%252F%252Farxiv.org%252Fabs%252F2405.20233.html

  424. Grounded language acquisition through the eyes and ears of a single child

  425. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2024-vong.pdf.html

  426. UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition

  427. https%253A%252F%252Farxiv.org%252Fabs%252F2311.15599%2523tencent.html

  428. ConvNets Match Vision Transformers at Scale

  429. https%253A%252F%252Farxiv.org%252Fabs%252F2310.16764%2523deepmind.html

  430. Interpret Vision Transformers as ConvNets with Dynamic Convolutions

  431. https%253A%252F%252Farxiv.org%252Fabs%252F2309.10713.html

  432. Rosetta Neurons: Mining the Common Units in a Model Zoo

  433. https%253A%252F%252Farxiv.org%252Fabs%252F2306.09346.html

  434. U-Net CNN in APL: Exploring Zero-Framework, Zero-Library Machine Learning

  435. https%253A%252F%252Fdl.acm.org%252Fdoi%252Fpdf%252F10.1145%252F3589246.3595371.html

  436. VanillaNet: the Power of Minimalism in Deep Learning

  437. https%253A%252F%252Farxiv.org%252Fabs%252F2305.12972.html

  438. ImageNet-Hard: The Hardest Images Remaining from a Study of the Power of Zoom and Spatial Biases in Image Classification

  439. https%253A%252F%252Farxiv.org%252Fabs%252F2304.05538.html

  440. Hierarchical Multi-Label Attribute Classification With Graph Convolutional Networks on Anime Illustration

  441. https%253A%252F%252Fieeexplore.ieee.org%252Fabstract%252Fdocument%252F10097719.html

  442. Pretraining Without Attention

  443. Junxiong Wang

  444. Albert Gu

  445. https://rush-nlp.com/

  446. https%253A%252F%252Farxiv.org%252Fabs%252F2212.10544.html

  447. What do Vision Transformers Learn? A Visual Exploration

  448. https%253A%252F%252Farxiv.org%252Fabs%252F2212.06727.html

  449. Simulated automated facial recognition systems as decision-aids in forensic face matching tasks

  450. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2022-carragher.pdf.html

  451. GCN-Based Multi-Modal Multi-Label Attribute Classification in Anime Illustration Using Domain-Specific Semantic Features

  452. %252Fdoc%252Fai%252Fanime%252Fdanbooru%252F2022-lan.pdf.html

  453. Understanding the Covariance Structure of Convolutional Filters

  454. https%253A%252F%252Farxiv.org%252Fabs%252F2210.03651.html

  455. Omnigrok: Grokking Beyond Algorithmic Data

  456. https%253A%252F%252Farxiv.org%252Fabs%252F2210.01117.html

  457. g.pt: Learning to Learn with Generative Models of Neural Network Checkpoints

  458. https%253A%252F%252Farxiv.org%252Fabs%252F2209.12892.html

  459. FastSiam: Resource-Efficient Self-supervised Learning on a Single GPU

  460. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2022-pototzky.pdf.html

  461. Reassessing hierarchical correspondences between brain and deep networks through direct interface

  462. https%253A%252F%252Fwww.science.org%252Fdoi%252F10.1126%252Fsciadv.abm2219.html

  463. RHO-LOSS: Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt

  464. https%253A%252F%252Farxiv.org%252Fabs%252F2206.07137.html

  465. BigVGAN: A Universal Neural Vocoder with Large-Scale Training

  466. https%253A%252F%252Farxiv.org%252Fabs%252F2206.04658%2523nvidia.html

  467. Scaling Up Your Kernels to 31×31: Revisiting Large Kernel Design in CNNs (RepLKNet)

  468. https%253A%252F%252Farxiv.org%252Fabs%252F2203.06717.html

  469. General Cyclical Training of Neural Networks

  470. https%253A%252F%252Farxiv.org%252Fabs%252F2202.08835.html

  471. Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet

  472. Matthias Bethge

  473. https%253A%252F%252Fopenreview.net%252Fforum%253Fid%253DSkfMWhAqYQ.html

  474. ConvMixer: Patches Are All You Need?

  475. https%253A%252F%252Farxiv.org%252Fabs%252F2201.09792.html

  476. ConvNeXt: A ConvNet for the 2020s

  477. Zhuang Liu’s Homepage

  478. https%253A%252F%252Farxiv.org%252Fabs%252F2201.03545%2523facebook.html

  479. AugMax: Adversarial Composition of Random Augmentations for Robust Training

  480. https%253A%252F%252Farxiv.org%252Fabs%252F2110.13771%2523nvidia.html

  481. TWIST: Self-Supervised Learning by Estimating Twin Class Distributions

  482. https%253A%252F%252Farxiv.org%252Fabs%252F2110.07402%2523bytedance.html

  483. Decoupled Contrastive Learning

  484. https%253A%252F%252Farxiv.org%252Fabs%252F2110.06848.html

  485. Supervision Exists Everywhere: A Data Efficient Contrastive Language-Image Pre-training Paradigm (DeCLIP)

  486. https%253A%252F%252Farxiv.org%252Fabs%252F2110.05208.html

  487. THINGSvision: A Python Toolbox for Streamlining the Extraction of Activations From Deep Neural Networks

  488. https%253A%252F%252Fwww.frontiersin.org%252Farticles%252F10.3389%252Ffninf.2021.679838%252Ffull.html

  489. A Battle of Network Structures: An Empirical Study of CNN, Transformer, and MLP

  490. https%253A%252F%252Farxiv.org%252Fabs%252F2108.13002%2523microsoft.html

  491. Do Vision Transformers See Like Convolutional Neural Networks?

  492. https%253A%252F%252Farxiv.org%252Fabs%252F2108.08810%2523google.html

  493. Neuroprosthesis for Decoding Speech in a Paralyzed Person with Anarthria

  494. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2021-moses.pdf.html

  495. Partial success in closing the gap between human and machine vision

  496. Robert Geirhos

  497. Matthias Bethge

  498. https%253A%252F%252Farxiv.org%252Fabs%252F2106.07411.html

  499. CoAtNet: Marrying Convolution and Attention for All Data Sizes

  500. Zihang Dai

  501. https%253A%252F%252Farxiv.org%252Fabs%252F2106.04803%2523google.html

  502. Effect of Pre-Training Scale on Intra/Inter-Domain Full and Few-Shot Transfer Learning for Natural and Medical X-Ray Chest Images

  503. https%253A%252F%252Farxiv.org%252Fabs%252F2106.00116.html

  504. Rethinking and Improving the Robustness of Image Style Transfer

  505. https%253A%252F%252Farxiv.org%252Fabs%252F2104.05623.html

  506. Rip van Winkle’s Razor, a Simple New Estimate for Adaptive Data Analysis

  507. https%253A%252F%252Fwww.offconvex.org%252F2021%252F04%252F07%252Fripvanwinkle%252F.html

  508. The surprising impact of mask-head architecture on novel class segmentation

  509. https%253A%252F%252Farxiv.org%252Fabs%252F2104.00613.html

  510. Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks

  511. https%253A%252F%252Farxiv.org%252Fabs%252F2103.14749.html

  512. ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases

  513. https%253A%252F%252Farxiv.org%252Fabs%252F2103.10697%2523facebook.html

  514. Learning from videos to understand the world

  515. Polina Kuznetsova

  516. https%253A%252F%252Fai.facebook.com%252Fblog%252Flearning-from-videos-to-understand-the-world%252F.html

  517. NFNet: High-Performance Large-Scale Image Recognition Without Normalization

  518. Karen Simonyan

  519. https%253A%252F%252Farxiv.org%252Fabs%252F2102.06171%2523deepmind.html

  520. Meta Pseudo Labels

  521. Zihang Dai

  522. https%253A%252F%252Farxiv.org%252Fabs%252F2003.10580%2523google.html

  523. Scaling down Deep Learning

  524. About Sam Greydanus

  525. https%253A%252F%252Fgreydanus.github.io%252F2020%252F12%252F01%252Fscaling-down%252F.html

  526. Very Deep VAEs Generalize Autoregressive Models and Can Outperform Them on Images

  527. https%253A%252F%252Farxiv.org%252Fabs%252F2011.10650%2523openai.html

  528. Sharpness-Aware Minimization (SAM) for Efficiently Improving Generalization

  529. Behnam Neyshabur

  530. https%253A%252F%252Farxiv.org%252Fabs%252F2010.01412%2523google.html

  531. Optimal Peanut Butter and Banana Sandwiches

  532. https%253A%252F%252Fwww.ethanrosenthal.com%252F2020%252F08%252F25%252Foptimal-peanut-butter-and-banana-sandwiches%252F.html

  533. Accuracy and Performance Comparison of Video Action Recognition Approaches

  534. https%253A%252F%252Farxiv.org%252Fabs%252F2008.09037.html

  535. SSD vs. YOLO for Detection of Outdoor Urban Advertising Panels under Multiple Variabilities

  536. https%253A%252F%252Fwww.mdpi.com%252F1424-8220%252F20%252F16%252F4587.html

  537. NVAE: A Deep Hierarchical Variational Autoencoder

  538. https%253A%252F%252Farxiv.org%252Fabs%252F2007.03898%2523nvidia.html

  539. YOLOv4: Optimal Speed and Accuracy of Object Detection

  540. https%253A%252F%252Farxiv.org%252Fabs%252F2004.10934.html

  541. Scaling Laws from the Data Manifold Dimension

  542. Jared Kaplan

  543. https%253A%252F%252Farxiv.org%252Fabs%252F2004.10802.html

  544. A Simple Framework for Contrastive Learning of Visual Representations

  545. https%253A%252F%252Farxiv.org%252Fabs%252F2002.05709%2523google.html

  546. Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving

  547. Stefano Ermon

  548. https%253A%252F%252Farxiv.org%252Fabs%252F2002.03629.html

  549. The Importance of Deconstruction

  550. Welcome

  551. https%253A%252F%252Fwww.youtube.com%252Fwatch%253Fv%253DkY2NHSKBi10.html

  552. Dynamic Convolution: Attention over Convolution Kernels

  553. https%253A%252F%252Farxiv.org%252Fabs%252F1912.03458%2523microsoft.html

  554. Deep Double Descent: We show that the double descent phenomenon occurs in CNNs, ResNets, and transformers: performance first improves, then gets worse, and then improves again with increasing model size, data size, or training time

  555. Preetum Nakkiran

  556. Yamini Bansal

  557. https%253A%252F%252Fopenai.com%252Fresearch%252Fdeep-double-descent.html

  558. Fantastic Generalization Measures and Where to Find Them

  559. Behnam Neyshabur

  560. https%253A%252F%252Farxiv.org%252Fabs%252F1912.02178.html

  561. Self-training with Noisy Student improves ImageNet classification

  562. https%253A%252F%252Farxiv.org%252Fabs%252F1911.04252%2523google.html

  563. DD-PPO: Learning Near-Perfect PointGoal Navigators from 2.5 Billion Frames

  564. https%253A%252F%252Farxiv.org%252Fabs%252F1911.00357%2523facebook.html

  565. Human-level performance in 3D multiplayer games with population-based reinforcement learning

  566. Guy Lever

  567. Koray Kavukcuoglu

  568. %252Fdoc%252Freinforcement-learning%252Fexploration%252F2019-jaderberg.pdf%2523deepmind.html

  569. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features

  570. https%253A%252F%252Farxiv.org%252Fabs%252F1905.04899.html

  571. Billion-scale semi-supervised learning for image classification

  572. https%253A%252F%252Farxiv.org%252Fabs%252F1905.00546%2523facebook.html

  573. A Recipe for Training Neural Networks

  574. https%253A%252F%252Fkarpathy.github.io%252F2019%252F04%252F25%252Frecipe%252F.html

  575. Detecting advertising on building façades with computer vision

  576. https%253A%252F%252Fwww.sciencedirect.com%252Fscience%252Farticle%252Fpii%252FS1877050919311299.html

  577. FloWaveNet: A Generative Flow for Raw Audio

  578. https%253A%252F%252Farxiv.org%252Fabs%252F1811.02155.html

  579. StreetNet: Preference Learning with Convolutional Neural Network on Urban Crime Perception

  580. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2018-fu.pdf.html

  581. Human-Like Playtesting with Deep Learning

  582. %252Fdoc%252Freinforcement-learning%252Fimitation-learning%252F2018-gudmundsson.pdf.html

  583. CurriculumNet: Weakly Supervised Learning from Large-Scale Web Images

  584. https%253A%252F%252Farxiv.org%252Fabs%252F1808.01097.html

  585. Exploring the Limits of Weakly Supervised Pretraining

  586. Ross Girshick

  587. Laurens Van Der Maaten

  588. https%253A%252F%252Farxiv.org%252Fabs%252F1805.00932%2523facebook.html

  589. YOLOv3: An Incremental Improvement

  590. https%253A%252F%252Farxiv.org%252Fabs%252F1804.02767.html

  591. Reptile: On First-Order Meta-Learning Algorithms

  592. John Schulman’s Homepage

  593. https%253A%252F%252Farxiv.org%252Fabs%252F1803.02999%2523openai.html

  594. Guess, check and fix: a phenomenology of improvisation in ‘neural’ painting

  595. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2018-choi.pdf.html

  596. China’s AI Advances Help Its Tech Industry, and State Security

  597. https%253A%252F%252Fwww.nytimes.com%252F2017%252F12%252F03%252Fbusiness%252Fchina-artificial-intelligence.html.html

  598. Learning to Play Chess with Minimal Lookahead and Deep Value Neural Networks

  599. %252Fdoc%252Freinforcement-learning%252Fchess%252F2017-sabatelli.pdf%2523page%253D3.html

  600. SMASH: One-Shot Model Architecture Search through HyperNetworks

  601. https%253A%252F%252Farxiv.org%252Fabs%252F1708.05344.html

  602. A Downsampled Variant of ImageNet as an Alternative to the CIFAR datasets

  603. Ilya Loshchilov

  604. Profile – Machine Learning Lab

  605. https%253A%252F%252Farxiv.org%252Fabs%252F1707.08819.html

  606. Towards Deep Learning Models Resistant to Adversarial Attacks

  607. Homepage: Aleksander Mądry

  608. https%253A%252F%252Farxiv.org%252Fabs%252F1706.06083.html

  609. A simple neural network module for relational reasoning

  610. https://sites.google.com/view/razp/home

  611. https%253A%252F%252Farxiv.org%252Fabs%252F1706.01427%2523deepmind.html

  612. What Makes a Good Image? Airbnb Demand Analytics Leveraging Interpretable Image Features

  613. Kannan Srinivasan—Analysis Group

  614. https%253A%252F%252Fpapers.ssrn.com%252Fsol3%252Fpapers.cfm%253Fabstract_id%253D2976021.html

  615. Scaling the Scattering Transform: Deep Hybrid Networks

  616. https%253A%252F%252Farxiv.org%252Fabs%252F1703.08961.html

  617. Mask R-CNN

  618. Ross Girshick

  619. https%253A%252F%252Farxiv.org%252Fabs%252F1703.06870%2523facebook.html

  620. Convolution Aware Initialization

  621. https%253A%252F%252Farxiv.org%252Fabs%252F1702.06295.html

  622. BrainNetCNN: Convolutional neural networks for brain networks; towards predicting neurodevelopment

  623. %252Fdoc%252Fpsychology%252Fneuroscience%252F2017-kawahara.pdf.html

  624. YOLO9000: Better, Faster, Stronger

  625. https%253A%252F%252Farxiv.org%252Fabs%252F1612.08242.html

  626. Self-critical Sequence Training for Image Captioning

  627. https%253A%252F%252Farxiv.org%252Fabs%252F1612.00563.html

  628. ResNeXt: Aggregated Residual Transformations for Deep Neural Networks

  629. Ross Girshick

  630. https%253A%252F%252Farxiv.org%252Fabs%252F1611.05431%2523facebook.html

  631. Clockwork Convnets for Video Semantic Segmentation

  632. https%253A%252F%252Farxiv.org%252Fabs%252F1608.03609.html

  633. Wide Residual Networks

  634. https%253A%252F%252Farxiv.org%252Fabs%252F1605.07146.html

  635. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks

  636. https%253A%252F%252Farxiv.org%252Fabs%252F1603.05279.html

  637. Microsoft researchers win ImageNet computer vision challenge

  638. https%253A%252F%252Fblogs.microsoft.com%252Fai%252Fmicrosoft-researchers-win-imagenet-computer-vision-challenge%252F.html

  639. The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition

  640. Jonathan Krause

  641. https%253A%252F%252Farxiv.org%252Fabs%252F1511.06789%2523google.html

  642. Learning Visual Features from Large Weakly Supervised Data

  643. Laurens Van Der Maaten

  644. https%253A%252F%252Farxiv.org%252Fabs%252F1511.02251%2523facebook.html

  645. Predicting and Understanding Urban Perception with Convolutional Neural Networks

  646. %252Fdoc%252Fai%252Fnn%252Fcnn%252F2015-porzi.pdf.html

  647. You Only Look Once: Unified, Real-Time Object Detection

  648. Ross Girshick

  649. https%253A%252F%252Farxiv.org%252Fabs%252F1506.02640.html

  650. Clothing-1M: Learning from Massive Noisy Labeled Data for Image Classification

  651. https%253A%252F%252Fopenaccess.thecvf.com%252Fcontent_cvpr_2015%252Fpapers%252FXiao_Learning_From_Massive_2015_CVPR_paper.pdf%2523baidu.html

  652. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

  653. Ross Girshick

  654. https%253A%252F%252Farxiv.org%252Fabs%252F1506.01497%2523microsoft.html

  655. Cyclical Learning Rates for Training Neural Networks

  656. https%253A%252F%252Farxiv.org%252Fabs%252F1506.01186.html

  657. Fast R-CNN

  658. Ross Girshick

  659. https%253A%252F%252Farxiv.org%252Fabs%252F1504.08083%2523microsoft.html

  660. R-CNN: Rich feature hierarchies for accurate object detection and semantic segmentation

  661. Ross Girshick

  662. https%253A%252F%252Farxiv.org%252Fabs%252F1311.2524.html

  663. DanNet: Flexible, High Performance Convolutional Neural Networks for Image Classification

  664. https%253A%252F%252Farxiv.org%252Fabs%252F1102.0183%2523schmidhuber.html