Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?
Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
HtmlRAG: HTML is Better Than Plain Text for Modeling Retrieved Knowledge in RAG Systems
Inference Scaling for Long-Context Retrieval Augmented Generation
Operational Advice for Dense and Sparse Retrievers: HNSW, Flat, or Inverted Indexes?
Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More?
OpenAI’s Colin Jarvis predicts "exponential" advancements in large language model capabilities during AI Summit London keynote
State Soup: In-Context Skill Learning, Retrieval and Mixing
Retrieval Head Mechanistically Explains Long-Context Factuality
Aligning LLM Agents by Learning Latent Preference from User Edits
Towards Generated Image Provenance Analysis Via Conceptual-Similar-Guided-SLIP Retrieval
FABLES: Evaluating faithfulness and content selection in book-length summarization
Online Adaptation of Language Models with a Memory of Amortized Contexts (MAC)
RNNs are not Transformers (Yet): The Key Bottleneck on In-context Retrieval
Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations (HSTU)
RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval
RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture
ReST meets ReAct: Self-Improvement for Multi-Step Reasoning LLM Agent
Look Before You Leap: A Universal Emergent Decomposition of Retrieval Tasks in Language Models
Retrieving Conditions from Reference Images for Diffusion Models
Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine
PEARL: Personalizing Large Language Model Writing Assistants with Generation-Calibrated Retrievers
In-Context Pretraining (ICP): Language Modeling Beyond Document Boundaries
SWE-bench: Can Language Models Resolve Real-World GitHub Issues?
FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation
RAVEN: In-Context Learning with Retrieval-Augmented Encoder-Decoder Language Models
Gzip versus bag-of-words for text classification with k-NN
LeanDojo: Theorem Proving with Retrieval-Augmented Language Models
TTT-NN: Test-Time Training on Nearest Neighbors for Large Language Models
Landmark Attention: Random-Access Infinite Context Length for Transformers
WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia
Long-Term Value of Exploration: Measurements, Findings and Algorithms
Unlimiformer: Long-Range Transformers with Unlimited Length Input
q2d: Turning Questions into Dialogs to Teach Models How to Search
CLaMP: Contrastive Language-Music Pre-training for Cross-Modal Symbolic Music Information Retrieval
Language Models Enable Simple Systems for Generating Structured Views of Heterogeneous Data Lakes
Shall We Pretrain Autoregressive Language Models with Retrieval? A Comprehensive Study
MaMMUT: A Simple Architecture for Joint Learning for MultiModal Tasks
Mitigating YouTube Recommendation Polarity using BERT and K-Means Clustering
ProofNet: Autoformalizing and Formally Proving Undergraduate-Level Mathematics
Not what you’ve signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection
Characterizing Attribution and Fluency Tradeoffs for Retrieval-Augmented Large Language Models
Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning
InPars-Light: Cost-Effective Unsupervised Training of Efficient Rankers
Precise Zero-Shot Dense Retrieval without Relevance Labels
Less is More: Parameter-Free Text Classification with Gzip
One Embedder, Any Task: Instruction-Finetuned Text Embeddings (INSTRUCTOR)
Text Embeddings by Weakly-Supervised Contrastive Pre-training
GENIUS: Sketch-based Language Model Pre-training via Extreme and Selective Masking for Text Generation and Augmentation
Large Language Models Struggle to Learn Long-Tail Knowledge
RARR: Attributed Text Generation via Post-hoc Research and Revision
Self-Ask: Measuring and Narrowing the Compositionality Gap in Language Models (Bamboogle)
ReAct: Synergizing Reasoning and Acting in Language Models
FiD-Light: Efficient and Effective Retrieval-Augmented Text Generation
Sparrow: Improving alignment of dialogue agents via targeted human judgements
Generate rather than Retrieve (GenRead): Large Language Models are Strong Context Generators
Vote-K: Selective Annotation Makes Language Models Better Few-Shot Learners
CorpusBrain: Pre-train a Generative Retrieval Model for Knowledge-Intensive Language Tasks
Text-Guided Synthesis of Artistic Images with Retrieval-Augmented Diffusion Models
NaturalProver: Grounded Mathematical Proof Generation with Language Models
Down and Across: Introducing Crossword-Solving as a New NLP Benchmark
RankGen: Improving Text Generation with Large Ranking Models
Retrieval-Augmented Diffusion Models: Semi-Parametric Neural Image Synthesis
Language Models that Seek for Knowledge: Modular Search & Generation for Dialogue and Prompt Completion
Unsupervised Vision-and-Language Pre-training via Retrieval-based Multi-Granular Alignment
Retrieval Augmented Classification for Long-Tail Visual Recognition
InPars: Data Augmentation for Information Retrieval using Large Language Models
Memory-assisted prompt editing to improve GPT-3 after deployment
A Thousand Words Are Worth More Than a Picture: Natural Language-Centric Outside-Knowledge Visual Question Answering
Contriever: Towards Unsupervised Dense Information Retrieval with Contrastive Learning
WebGPT: Browser-assisted question-answering with human feedback
WebGPT: Improving the factual accuracy of language models through web browsing
You Only Need One Model for Open-domain Question Answering
Improving language models by retrieving from trillions of tokens
Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention
Scaling Law for Recommendation Models: Towards General-purpose User Representations
SPANN: Highly-efficient Billion-scale Approximate Nearest Neighbor Search
HTCN: Harmonious Text Colorization Network for Visual-Textual Presentation Design
CLOOB: Modern Hopfield Networks with InfoLOOB Outperform CLIP
One Loss for All: Deep Hashing with a Single Cosine Similarity based Learning Objective
SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval
MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance Detection
EfficientCLIP: Efficient Cross-Modal Pre-training by Ensemble Confident Learning and Language Modeling
Contrastive Language-Image Pre-training for the Italian Language
Sentence-T5: Scalable Sentence Encoders from Pre-trained Text-to-Text Models
Billion-Scale Pretraining with Vision Transformers for Multi-Task Visual Representations
MuSiQue: Multi-hop Questions via Single-hop Question Composition
SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking
A Multi-Level Attention Model for Evidence-Based Fact Checking
Towards mental time travel: a hierarchical memory for reinforcement learning agents
RetGen: A Joint framework for Retrieval and Grounded Text Generation Modeling
Not All Memories are Created Equal: Learning to Forget by Expiring
Rethinking Search: Making Domain Experts out of Dilettantes
SimCSE: Simple Contrastive Learning of Sentence Embeddings
BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models
Retrieval Augmentation Reduces Hallucination in Conversation
TSDAE: Using Transformer-based Sequential Denoising Autoencoder for Unsupervised Sentence Embedding Learning
NaturalProofs: Mathematical Theorem Proving in Natural Language
China’s GPT-3? BAAI Introduces Superscale Intelligence Model ‘Wu Dao 1.0’: The Beijing Academy of Artificial Intelligence (BAAI) releases Wu Dao 1.0, China’s first large-scale pretraining model.
Get Your Vitamin C! Robust Fact Verification with Contrastive Evidence (VitaminC)
ALIGN: Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers
Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup
Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps
Current Limitations of Language Models: What You Need is Retrieval
Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering
M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training
System for searching illustrations of anime characters focusing on degrees of character attributes
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
Dense Passage Retrieval for Open-Domain Question Answering
Learning to Scale Multilingual Representations for Vision-Language Tasks
How Much Knowledge Can You Pack Into the Parameters of a Language Model?
REALM: Integrating Retrieval into Language Representation Models
SimpleShot: Revisiting Nearest-Neighbor Classification for Few-Shot Learning
Generalization through Memorization: Nearest Neighbor Language Models
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge
LIGHT: Learning to Speak and Act in a Fantasy Text Adventure Game
Top-K Off-Policy Correction for a REINFORCE Recommender System
FEVER: a large-scale dataset for Fact Extraction and VERification
Towards Deep Modeling of Music Semantics using EEG Regularizers
Learning to Organize Knowledge and Answer Questions with N-Gram Machines
Seq2SQL: Generating Structured Queries from Natural Language using Reinforcement Learning
Bolt: Accelerated Data Mining with Fast Vector Compression
Ask the Right Questions: Active Question Reformulation with Reinforcement Learning
Get To The Point: Summarization with Pointer-Generator Networks
Scaling Memory-Augmented Neural Networks with Sparse Reads and Writes
Improving Information Extraction by Acquiring External Evidence with Reinforcement Learning
PlaNet—Photo Geolocation with Convolutional Neural Networks
Illustration2Vec: a semantic vector representation of illustrations
Learning to Win by Reading Manuals in a Monte-Carlo Framework
This Week’s Citation Classic: Nearest Neighbor Pattern Classification
ANN-Benchmarks Is a Benchmarking Environment for Approximate Nearest Neighbor Algorithms Search. This Website Contains the Current Benchmarking Results. Please Visit Https://github.com/erikbern/ann-Benchmarks/ to Get an Overview over Evaluated Data Sets and Algorithms. Make a Pull Request on Github to Add Your Own Code or Improvements to the Benchmarking System.
Find Anything Blazingly Fast With Google's Vector Search Technology
This Anime Does Not Exist, Search: This Notebook Uses the Precomputed CLIP Feature Vectors for 100k Images from TADNE
PaddlePaddle/RocketQA: 🚀 RocketQA, Dense Retrieval for Information Retrieval and Question Answering, including Both Chinese and English State-Of-The-Art Models.
Building a Vector Database in 2GB for 36 Million Wikipedia Passages
The Super Effectiveness of Pokémon Embeddings Using Only Raw JSON and Images
European Parliament Revolutionizes Archive Access With Claude AI
Azure AI Milestone: Microsoft KEAR Surpasses Human Performance on CommonsenseQA Benchmark
Turing Bletchley: A Universal Image Language Representation Model by Microsoft
Here Are 120K 𝑤 Samples from @AydaoAI’s Large Anime Model (aka TADNE) Clustered into a Set of 256 Centroids. 𝘸𝘢𝘵𝘤𝘩 𝘪𝘵 𝘴𝘩𝘪𝘯𝘦
2023-girdhar-figure1-imagebindsjointembeddingspacenablesemergentmultimodalcapabilitieslikeembeddingarithmeticoraudio2imagegeneration.jpg
2023-girdhar-figure5-objectdetectioninimageswithaudioqueriesinimagebindrequiringnoretraining.png
2022-gao-figure1-hydearchitecturediagramofhallucinatingananswerandthenlookingupsimilardocumentstousetogeneratearealanswer.png
2022-press-figure5-selfaskplusgooglesearchengine-innermonologueforsearchingtheinternettoanswermultihopquestions.png
2022-press-table1-selfaskplusgooglesearchengine-innermonologueforsearchingtheinternettoanswermultihopquestions-benchmarkperformance.jpg
https://about.sourcegraph.com/blog/cheating-is-all-you-need
https://ai.meta.com/blog/next-generation-meta-training-inference-accelerator-AI-MTIA/
https://aimd.app/blog/2024-01-16-using-ai-to-overengineer-404-pages
https://ashvardanian.com/posts/python-c-assembly-comparison/
https://blog.pgvecto.rs/my-binary-vector-search-is-better-than-your-fp32-vectors
https://cookbook.openai.com/examples/tag_caption_images_with_gpt4v
https://economistwritingeveryday.com/2024/01/07/using-phind-for-academic-references/
https://engineering.fb.com/2017/03/29/data-infrastructure/faiss-a-library-for-efficient-similarity-search/
https://every.to/chain-of-thought/gpt-4-is-a-reasoning-engine
https://every.to/chain-of-thought/i-spent-a-week-with-gemini-pro-1-5-it-s-fantastic
https://openai.com/blog/introducing-text-and-code-embeddings/
https://openai.com/index/new-embedding-models-and-api-updates/
https://platform.openai.com/docs/guides/embeddings/use-cases
https://python.langchain.com/v0.1/docs/modules/data_connection/retrievers/multi_vector/
https://research.google/blog/kelm-integrating-knowledge-graphs-with-language-model-pre-training-corpora/
https://searchengineland.com/how-google-search-ranking-works-pandu-nayak-435395#h-navboost-system-a-k-a-glue
https://simonwillison.net/2024/Apr/17/ai-for-data-journalism/
https://til.simonwillison.net/llms/claude-hacker-news-themes
https://tyleransom.substack.com/p/using-llms-to-fuzzy-merge
https://www.askviable.com/blog/why-we-chose-gpt-3-embeddings-for-the-clustering-behind-our-feedback-reports
https://www.oreilly.com/radar/what-we-learned-from-a-year-of-building-with-llms-part-i/
https://www.reddit.com/r/ChatGPT/comments/12a0ajb/i_gave_gpt4_persistent_memory_and_the_ability_to/
https://www.reddit.com/r/MachineLearning/comments/117yw1w/d_maybe_a_new_prompt_injection_method_against/
https://www.reddit.com/r/MachineLearning/comments/1fyb9jj/p_model2vec_distill_a_small_fast_model_from_any/
Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More?
https%253A%252F%252Farxiv.org%252Fabs%252F2406.13121%2523google.html
OpenAI’s Colin Jarvis predicts "exponential" advancements in large language model capabilities during AI Summit London keynote
https%253A%252F%252Faibusiness.com%252Fnlp%252Fopenai-chief-architect-predicts-huge-large-language-model-leaps.html
Retrieval Head Mechanistically Explains Long-Context Factuality
https%253A%252F%252Farxiv.org%252Fabs%252F2403.18802%2523deepmind.html
Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations (HSTU)
https%253A%252F%252Farxiv.org%252Fabs%252F2402.17152%2523facebook.html
RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture
https%253A%252F%252Farxiv.org%252Fabs%252F2401.08406%2523microsoft.html
Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine
https%253A%252F%252Farxiv.org%252Fabs%252F2311.16452%2523microsoft.html
FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation
https%253A%252F%252Farxiv.org%252Fabs%252F2310.03214%2523google.html
LeanDojo: Theorem Proving with Retrieval-Augmented Language Models
TTT-NN: Test-Time Training on Nearest Neighbors for Large Language Models
Landmark Attention: Random-Access Infinite Context Length for Transformers
https%253A%252F%252Farxiv.org%252Fabs%252F2305.05665%2523facebook.html
Unlimiformer: Long-Range Transformers with Unlimited Length Input
q2d: Turning Questions into Dialogs to Teach Models How to Search
https%253A%252F%252Farxiv.org%252Fabs%252F2304.14318%2523google.html
Shall We Pretrain Autoregressive Language Models with Retrieval? A Comprehensive Study
https%253A%252F%252Farxiv.org%252Fabs%252F2304.06762%2523nvidia.html
ProofNet: Autoformalizing and Formally Proving Undergraduate-Level Mathematics
Not what you’ve signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection
Precise Zero-Shot Dense Retrieval without Relevance Labels
Less is More: Parameter-Free Text Classification with Gzip
One Embedder, Any Task: Instruction-Finetuned Text Embeddings (INSTRUCTOR)
Text Embeddings by Weakly-Supervised Contrastive Pre-training
https%253A%252F%252Farxiv.org%252Fabs%252F2212.03533%2523microsoft.html
https%253A%252F%252Farxiv.org%252Fabs%252F2212.01349%2523facebook.html
https%253A%252F%252Farxiv.org%252Fabs%252F2211.12561%2523facebook.html
Large Language Models Struggle to Learn Long-Tail Knowledge
RARR: Attributed Text Generation via Post-hoc Research and Revision
https%253A%252F%252Farxiv.org%252Fabs%252F2210.08726%2523google.html
Self-Ask: Measuring and Narrowing the Compositionality Gap in Language Models (Bamboogle)
https%253A%252F%252Farxiv.org%252Fabs%252F2210.03350%2523allen.html
Vote-K: Selective Annotation Makes Language Models Better Few-Shot Learners
https%253A%252F%252Farxiv.org%252Fabs%252F2207.06300%2523ibm.html
https%253A%252F%252Farxiv.org%252Fabs%252F2206.05314%2523deepmind.html
https%253A%252F%252Fopenreview.net%252Fforum%253Fid%253D0ZbPmmB61g%2523google.html
NaturalProver: Grounded Mathematical Proof Generation with Language Models
https%253A%252F%252Farxiv.org%252Fabs%252F2205.12910%2523allen.html
https%253A%252F%252Farxiv.org%252Fabs%252F2205.05131%2523google.html
Language Models that Seek for Knowledge: Modular Search & Generation for Dialogue and Prompt Completion
https%253A%252F%252Farxiv.org%252Fabs%252F2203.13224%2523facebook.html
https%253A%252F%252Farxiv.org%252Fabs%252F2201.10005%2523openai.html
Contriever: Towards Unsupervised Dense Information Retrieval with Contrastive Learning
https%253A%252F%252Farxiv.org%252Fabs%252F2112.09118%2523facebook.html
WebGPT: Browser-assisted question-answering with human feedback
https%253A%252F%252Farxiv.org%252Fabs%252F2112.09332%2523openai.html
WebGPT: Improving the factual accuracy of language models through web browsing
https%253A%252F%252Fopenai.com%252Fresearch%252Fwebgpt.html
https%253A%252F%252Farxiv.org%252Fabs%252F2112.07899%2523google.html
Improving language models by retrieving from trillions of tokens
https%253A%252F%252Farxiv.org%252Fabs%252F2112.04426%2523deepmind.html
https%253A%252F%252Farxiv.org%252Fabs%252F2111.11432%2523microsoft.html
https%253A%252F%252Farxiv.org%252Fabs%252F2111.07991%2523google.html
Scaling Law for Recommendation Models: Towards General-purpose User Representations
https%253A%252F%252Farxiv.org%252Fabs%252F2203.08913%2523google.html
CLOOB: Modern Hopfield Networks with InfoLOOB Outperform CLIP
https%253A%252F%252Fopenreview.net%252Fforum%253Fid%253Dqw674L9PfQE.html
Sentence-T5: Scalable Sentence Encoders from Pre-trained Text-to-Text Models
https%253A%252F%252Farxiv.org%252Fabs%252F2108.08877%2523google.html
https%253A%252F%252Farxiv.org%252Fabs%252F2107.07566%2523facebook.html
Retrieval Augmentation Reduces Hallucination in Conversation
https%253A%252F%252Farxiv.org%252Fabs%252F2104.07567%2523facebook.html
China’s GPT-3? BAAI Introduces Superscale Intelligence Model ‘Wu Dao 1.0’: The Beijing Academy of Artificial Intelligence (BAAI) releases Wu Dao 1.0, China’s first large-scale pretraining model.
https%253A%252F%252Fsyncedreview.com%252F2021%252F03%252F23%252Fchinas-gpt-3-baai-introduces-superscale-intelligence-model-wu-dao-1-0%252F%2523baai.html
ALIGN: Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
https%253A%252F%252Farxiv.org%252Fabs%252F2102.05918%2523google.html
https%253A%252F%252Fwww.youtube.com%252Fwatch%253Fv%253DkY2NHSKBi10.html
This Week’s Citation Classic: Nearest Neighbor Pattern Classification
%252Fdoc%252Fai%252Fnn%252Fretrieval%252F1982-cover.pdf.html
Wikipedia Bibliography: