Bibliography:

  1. ‘AI’ tag

  2. ‘Highleyman’s AI’ tag

  3. Fully-Connected Neural Nets

  4. Weather and My Productivity

  5. Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting & Beyond

  6. Questionable practices in machine learning

  7. Probing the Decision Boundaries of In-context Learning in Large Language Models

  8. Attention as an RNN

  9. The harms of class imbalance corrections for machine learning based prediction models: a simulation study

  10. Many-Shot In-Context Learning

  11. From Words to Numbers: Your Large Language Model Is Secretly A Capable Regressor When Given In-Context Examples

  12. Chronos: Learning the Language of Time Series

  13. StructLM: Towards Building Generalist Models for Structured Knowledge Grounding

  14. Why do Random Forests Work? Understanding Tree Ensembles as Self-Regularizing Adaptive Smoothers

  15. Illusory generalizability of clinical prediction models

  16. Attention versus Contrastive Learning of Tabular Data—A Data-centric Benchmarking

  17. TabLib: A Dataset of 627M Tables with Context

  18. Unambiguous discrimination of all 20 proteinogenic amino acids and their modifications by nanopore

  19. Generating and Imputing Tabular Data via Diffusion and Flow-based Gradient-Boosted Trees

  20. Generating tabular datasets under differential privacy

  21. TableGPT: Towards Unifying Tables, Nature Language and Commands into One GPT

  22. Language models are weak learners

  23. RGD: Stochastic Re-weighted Gradient Descent via Distributionally Robust Optimization

  24. Large Language Models are Few-Shot Health Learners

  25. Deep Learning based Forecasting: a case study from the online fashion industry

  26. Language Models Enable Simple Systems for Generating Structured Views of Heterogeneous Data Lakes

  27. TSMixer: An All-MLP Architecture for Time Series Forecasting

  28. Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning

  29. Fast semi-supervised self-training algorithm based on data editing

  30. Table-To-Text generation and pre-training with TabT5

  31. Language Models are Realistic Tabular Data Generators

  32. Forecasting with trees

  33. Why do tree-based models still outperform deep learning on tabular data?

  34. Revisiting Pretraining Objectives for Tabular Deep Learning

  35. TabPFN: Meta-Learning a Real-Time Tabular AutoML Method For Small Data

  36. Transfer Learning with Deep Tabular Models

  37. Hopular: Modern Hopfield Networks for Tabular Data

  38. Predicting romantic interest during early relationship development: A preregistered investigation using machine learning

  39. On Embeddings for Numerical Features in Tabular Deep Learning

  40. To SMOTE, or not to SMOTE?

  41. M5 accuracy competition: Results, findings, and conclusions

  42. The GatedTabTransformer: An enhanced deep learning architecture for tabular modeling

  43. PFNs: Transformers Can Do Bayesian Inference

  44. DANets: Deep Abstract Networks for Tabular Data Classification and Regression

  45. Deep Neural Networks and Tabular Data: A Survey

  46. An Unsupervised Model for Identifying and Characterizing Dark Web Forums

  47. TAPEX: Table Pre-training via Learning a Neural SQL Executor

  48. ARM-Net: Adaptive Relation Modeling Network for Structured Data

  49. Decision tree heuristics can fail, even in the smoothed setting

  50. SCARF: Self-Supervised Contrastive Learning using Random Feature Corruption

  51. Revisiting Deep Learning Models for Tabular Data

  52. The Epic Sepsis Model Falls Short—The Importance of External Validation

  53. Well-tuned Simple Nets Excel on Tabular Datasets

  54. Tabular Data: Deep Learning is Not All You Need

  55. Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning

  56. SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training

  57. Intelligence and General Psychopathology in the Vietnam Experience Study: A Closer Look

  58. Converting tabular data into images for deep learning with convolutional neural networks

  59. External Validation of a Widely Implemented Proprietary Sepsis Prediction Model in Hospitalized Patients

  60. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting

  61. TabTransformer: Tabular Data Modeling Using Contextual Embeddings

  62. Engineering In-place (Shared-memory) Sorting Algorithms

  63. Kaggle forecasting competitions: An overlooked learning opportunity

  64. TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data

  65. Neural Additive Models: Interpretable Machine Learning with Neural Nets

  66. TAPAS: Weakly Supervised Table Parsing via Pre-training

  67. A Market in Dream: the Rapid Development of Anonymous Cybercrime

  68. VIME: Extending the Success of Self-supervised and Semi-supervised Learning to Tabular Domain

  69. Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods

  70. The Bouncer Problem: Challenges to Remote Explainability

  71. OHAC: Online Hierarchical Clustering Approximations

  72. LightGBM: A Highly Efficient Gradient Boosting Decision Tree

  73. TabNet: Attentive Interpretable Tabular Learning

  74. 3D human pose estimation via human structure-aware fully connected network

  75. ID3 Learns Juntas for Smoothed Product Distributions

  76. Behavioral Patterns in Smartphone Usage Predict Big Five Personality Traits

  77. Asymptotic learning curves of kernel methods: empirical data versus Teacher-Student paradigm

  78. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting

  79. SuperTML: Two-Dimensional Word Embedding for the Precognition on Structured Tabular Data

  80. Fairwashing: the risk of rationalization

  81. Tweedie Gradient Boosting for Extremely Unbalanced Zero-inflated Data

  82. Neural Arithmetic Logic Units

  83. Learning and Memorization

  84. Large-scale comparison of machine learning methods for drug target prediction on ChEMBL

  85. Repurposing High-Throughput Image Assays Enables Biological Activity Prediction for Drug Discovery

  86. Improving palliative care with deep learning

  87. Using Posters to Recommend Anime and Mangas in a Cold-Start Scenario

  88. Neural Collaborative Filtering

  89. OpenML Benchmarking Suites

  90. CatBoost: unbiased boosting with categorical features

  91. Resource-Efficient Machine Learning in 2 KB RAM for the Internet of Things

  92. XGBoost: A Scalable Tree Boosting System

  93. "Why Should I Trust You?": Explaining the Predictions of Any Classifier

  94. The MovieLens Datasets: History and Context

  95. Planning as satisfiability: Heuristics

  96. Leakage in Data Mining: Formulation, Detection, and Avoidance

  97. d2c8847e4ba0d751e74f1d15507700d4feb1f238.pdf

  98. Random survival forests

  99. Tree Induction vs. Logistic Regression: A Learning-Curve Analysis

  100. A Survey of Methods for Scaling Up Inductive Algorithms

  101. On the Boosting Ability of Top-Down Decision Tree Learning Algorithms

  102. On The Effect of Data Set Size on Bias And Variance in Classification Learning

  103. The Effects of Training Set Size on Decision Tree Complexity

  104. Scaling up the accuracy of Naive-Bayes classifiers: a decision-tree hybrid

  105. Stupid Data Miner Tricks: Overfitting the S&P 500

  106. The MONK’s Problems-A Performance Comparison of Different Learning Algorithms

  107. Symbolic and neural learning algorithms: An experimental comparison

  108. A Meta-Analysis of Overfitting in Machine Learning

  109. Statistical Modeling: The Two Cultures

  110. How Good Are LLMs at Doing ML on an Unknown Dataset?

  111. design#future-tag-features

    [Transclude the forward-link's context]

  112. 2022-grinsztajn-figure9-treesvsneuralnetson4classificationtasksusingnumericalfeaturesonmediumvslargedatasets.png

  113. https://github.com/siboehm/lleaves

  114. https://linktransformer.github.io/

  115. https://siboehm.com/articles/21/lleaves

  116. https://spectrum.ieee.org/its-too-easy-to-hide-bias-in-deeplearning-systems

  117. https://www.maskaravivek.com/post/gan-synthetic-data-generation/

  118. 9e486948ca931eeb1d863351c29e0470f2c0b695.html

  119. https://www.oneusefulthing.org/p/it-is-starting-to-get-strange

  120. https://www.reddit.com/r/Anki/comments/1c29775/fsrs_is_one_of_the_most_accurate_spaced/

  121. 00c123cd325e6d1cfdbee6633abfc9baddb20049.html

  122. https://www.thelancet.com/journals/lanhl/article/PIIS2666-7568(23)00189-7/fulltext

  123. https://x.com/emollick/status/1652170706312896512

  124. https://x.com/shinboson/status/1769231110691500140

  125. Probing the Decision Boundaries of In-context Learning in Large Language Models

  126. Aditya Grover

  127. https%253A%252F%252Farxiv.org%252Fabs%252F2406.11233.html

  128. From Words to Numbers: Your Large Language Model Is Secretly A Capable Regressor When Given In-Context Examples

  129. https%253A%252F%252Farxiv.org%252Fabs%252F2404.07544.html

  130. StructLM: Towards Building Generalist Models for Structured Knowledge Grounding

  131. https%253A%252F%252Farxiv.org%252Fabs%252F2402.16671.html

  132. RGD: Stochastic Re-weighted Gradient Descent via Distributionally Robust Optimization

  133. https%253A%252F%252Farxiv.org%252Fabs%252F2306.09222%2523google.html

  134. TSMixer: An All-MLP Architecture for Time Series Forecasting

  135. https%253A%252F%252Farxiv.org%252Fabs%252F2303.06053%2523google.html

  136. Forecasting with trees

  137. https%253A%252F%252Fwww.sciencedirect.com%252Fscience%252Farticle%252Fpii%252FS0169207021001679.html

  138. TabPFN: Meta-Learning a Real-Time Tabular AutoML Method For Small Data

  139. Profile – Machine Learning Lab

  140. https%253A%252F%252Farxiv.org%252Fabs%252F2207.01848.html

  141. Predicting romantic interest during early relationship development: A preregistered investigation using machine learning

  142. %252Fdoc%252Fsociology%252Ftechnology%252F2022-eastwick.pdf.html

  143. M5 accuracy competition: Results, findings, and conclusions

  144. https%253A%252F%252Fwww.sciencedirect.com%252Fscience%252Farticle%252Fpii%252FS0169207021001874.html

  145. PFNs: Transformers Can Do Bayesian Inference

  146. Profile – Machine Learning Lab

  147. https%253A%252F%252Farxiv.org%252Fabs%252F2112.10510.html

  148. Asymptotic learning curves of kernel methods: empirical data versus Teacher-Student paradigm

  149. https%253A%252F%252Farxiv.org%252Fabs%252F1905.10843.html

  150. Tree Induction vs. Logistic Regression: A Learning-Curve Analysis

  151. %252Fdoc%252Fai%252Fscaling%252F2003-perlich.pdf.html

  152. On the Boosting Ability of Top-Down Decision Tree Learning Algorithms

  153. https%253A%252F%252Fwww.sciencedirect.com%252Fscience%252Farticle%252Fpii%252FS0022000097915439.html