“‘Instruct-Tuning LLMs’ Tag”,2021-04-18 (; backlinks):
![]()
Bibliography for tag
ai/nn/transformer/gpt/instruction-tuning, most recent first: 5 related tags, 79 annotations, & 25 links (parent).
- See Also
- Links
- “SANA: Efficient High-Resolution Image Synthesis With Linear Diffusion Transformers”, et al 2024
- “Instruction Following without Instruction Tuning”, et al 2024
- “Hermes 3 Technical Report”, et al 2024
- “State Soup: In-Context Skill Learning, Retrieval and Mixing”, et al 2024
- “Auto Evol-Instruct: Automatic Instruction Evolving for Large Language Models”, et al 2024
- “Instruction Modeling: Instruction Tuning With Loss Over Instructions”, et al 2024
- “The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions”, et al 2024
- “Best Practices and Lessons Learned on Synthetic Data for Language Models”, et al 2024
- “RecurrentGemma: Moving Past Transformers for Efficient Open Language Models”, et al 2024
- “Length-Controlled AlpacaEval: A Simple Way to Debias Automatic Evaluators”, et al 2024
- “COIG-CQIA: Quality Is All You Need for Chinese Instruction Fine-Tuning”, et al 2024
- “MetaAligner: Conditional Weak-To-Strong Correction for Generalizable Multi-Objective Alignment of Language Models”, et al 2024
- “Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models”, et al 2024
- “StructLM: Towards Building Generalist Models for Structured Knowledge Grounding”, et al 2024
- “How to Train Data-Efficient LLMs”, et al 2024
- “Rephrasing the Web (WARP): A Recipe for Compute and Data-Efficient Language Modeling”, et al 2024
- “WaveCoder: Widespread And Versatile Enhanced Instruction Tuning With Refined Data Generation”, et al 2023
- “VILA: On Pre-Training for Visual Language Models”, et al 2023
- “Instruction-Tuning Aligns LLMs to the Human Brain”, et al 2023
- “R-Tuning: Teaching Large Language Models to Refuse Unknown Questions”, et al 2023
- “When ‘A Helpful Assistant’ Is Not Really Helpful: Personas in System Prompts Do Not Improve Performances of Large Language Models”, et al 2023
- “Language Models Are Super Mario (DARE): Absorbing Abilities from Homologous Models As a Free Lunch”, et al 2023
- “ChipNeMo: Domain-Adapted LLMs for Chip Design”, et al 2023
- “Mistral-7B”, et al 2023
- “LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models”, et al 2023
- “LLaVA-1.5: Improved Baselines With Visual Instruction Tuning”, et al 2023
- “UltraFeedback: Boosting Language Models With High-Quality Feedback”, et al 2023
- “AceGPT, Localizing Large Language Models in Arabic”, et al 2023
- “Can Programming Languages Boost Each Other via Instruction Tuning?”, et al 2023
- “DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection for Conversational AI”, et al 2023
- “LLaMA-2: Open Foundation and Fine-Tuned Chat Models”, et al 2023
- “AlpaGasus: Training A Better Alpaca With Fewer Data”, et al 2023
- “Instruction Mining: High-Quality Instruction Data Selection for Large Language Models”, et al 2023
- “Lost in the Middle: How Language Models Use Long Contexts”, et al 2023
- “On the Exploitability of Instruction Tuning”, et al 2023
- “ChessGPT: Bridging Policy Learning and Language Modeling”, et al 2023
- “Dr. LLaMa: Improving Small Language Models in Domain-Specific QA via Generative Data Augmentation”, et al 2023
- “SELF-ALIGN: Principle-Driven Self-Alignment of Language Models from Scratch With Minimal Human Supervision”, et al 2023
- “Distilling Step-By-Step! Outperforming Larger Language Models With Less Training Data and Smaller Model Sizes”, et al 2023
- “LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions”, et al 2023
- “WizardLM: Empowering Large Language Models to Follow Complex Instructions”, et al 2023
- “TANGO: Text-To-Audio Generation Using Instruction-Tuned LLM and Latent Diffusion Model”, et al 2023
- “Phoenix: Democratizing ChatGPT across Languages”, et al 2023
- “How Well Do Large Language Models Perform in Arithmetic Tasks?”, et al 2023
- “Larger Language Models Do In-Context Learning Differently”, et al 2023
- “LLaMa-1: Open and Efficient Foundation Language Models”, et al 2023
- “How Does In-Context Learning Help Prompt Tuning?”, et al 2023
- “Med-PaLM: Large Language Models Encode Clinical Knowledge”, et al 2022
- “Self-Instruct: Aligning Language Models With Self-Generated Instructions”, et al 2022
- “Unnatural Instructions: Tuning Language Models With (Almost) No Human Labor”, et al 2022
- “One Embedder, Any Task: Instruction-Finetuned Text Embeddings (INSTRUCTOR)”, et al 2022
- “HALIE: Evaluating Human-Language Model Interaction”, et al 2022
- “BLOOMZ/mT0: Crosslingual Generalization through Multitask Finetuning”, et al 2022
- “Help Me Write a Poem: Instruction Tuning As a Vehicle for Collaborative Poetry Writing (CoPoet)”, et al 2022
- “FLAN: Scaling Instruction-Finetuned Language Models”, et al 2022
- “Language Models Are Multilingual Chain-Of-Thought Reasoners”, et al 2022
- “LINGUIST: Language Model Instruction Tuning to Generate Annotated Utterances for Intent Classification and Slot Tagging”, et al 2022
- “Z-Code++: A Pre-Trained Language Model Optimized for Abstractive Summarization”, et al 2022
- “Few-Shot Adaptation Works With UnpredicTable Data”, et al 2022
- “RST: ReStructured Pre-Training”, 2022
- “InstructDial: Improving Zero and Few-Shot Generalization in Dialogue through Instruction Tuning”, et al 2022
- “CT0: Fine-Tuned Language Models Are Continual Learners”, et al 2022
- “Tk-Instruct: Benchmarking Generalization via In-Context Instructions on 1,600+ Language Tasks”, et al 2022
- “What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?”, et al 2022
- “UnifiedQA-V2: Stronger Generalization via Broader Cross-Format Training”, et al 2022
- “Reasoning Like Program Executors”, et al 2022
- “ZeroPrompt: Scaling Prompt-Based Pretraining to 1,000 Tasks Improves Zero-Shot Generalization”, et al 2022
- “ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning”, et al 2021
- “MetaICL: Learning to Learn In Context”, et al 2021
- “T0: Multitask Prompted Training Enables Zero-Shot Task Generalization”, et al 2021
- “FLAN: Finetuned Language Models Are Zero-Shot Learners”, et al 2021
- “Cross-Task Generalization via Natural Language Crowdsourcing Instructions”, et al 2021
- “CrossFit: A Few-Shot Learning Challenge for Cross-Task Generalization in NLP”, et al 2021
- “Adapting Language Models for Zero-Shot Learning by Meta-Tuning on Dataset and Prompt Collections”, et al 2021
- “Muppet: Massive Multi-Task Representations With Pre-Finetuning”, et al 2021
- “UnifiedQA: Crossing Format Boundaries With a Single QA System”, et al 2020
- “The Natural Language Decathlon: Multitask Learning As Question Answering”, et al 2018
- “No Robots: Look Ma, an Instruction Dataset That Wasn’t Generated by GPTs!”, Hugging2024
- “The RetroInstruct Guide To Synthetic Text Data”, 2024
- Miscellaneous
- Bibliography