- See Also
-
Links
- “Interactive-Chain-Prompting (INTERCPT): Ambiguity Resolution for Crosslingual Conditional Generation With Interaction”, Et Al 2023
- “U-PaLM: Transcending Scaling Laws With 0.1% Extra Compute”, Et Al 2022
- “Challenging BIG-Bench Tasks (BBH) and Whether Chain-of-Thought Can Solve Them”, Et Al 2022
- “Language Models Are Multilingual Chain-of-Thought Reasoners”, Et Al 2022
- “Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit”, Et Al 2022
- “Emergent Abilities of Large Language Models”, Et Al 2022
- “Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models”, Et Al 2022
- “Data Distributional Properties Drive Emergent Few-Shot Learning in Transformers”, Et Al 2022
- “PaLM: Scaling Language Modeling With Pathways”, Et Al 2022
- “In-context Learning and Induction Heads”, Et Al 2022
- “Predictability and Surprise in Large Generative Models”, Et Al 2022
- “The Effects of Reward Misspecification: Mapping and Mitigating Misaligned Models”, Et Al 2022
- “A Mathematical Framework for Transformer Circuits”, Et Al 2021
- “Scaling Language Models: Methods, Analysis & Insights from Training Gopher”, Et Al 2021
- “A General Language Assistant As a Laboratory for Alignment”, Et Al 2021
- “Mapping Language Models to Grounded Conceptual Spaces”, 2021
- “Program Synthesis With Large Language Models”, Et Al 2021
- “MMLU: Measuring Massive Multitask Language Understanding”, Et Al 2020
- “GPT-3: Language Models Are Few-Shot Learners”, Et Al 2020
- “Emergence in Cognitive Science”, 2010
- “The Phase Transition In Human Cognition § Phase Transitions in Language Processing”, Et Al 2009 (page 13)
- Wikipedia
- Miscellaneous
- Link Bibliography
See Also
Links
“Interactive-Chain-Prompting (INTERCPT): Ambiguity Resolution for Crosslingual Conditional Generation With Interaction”, Et Al 2023
“Interactive-Chain-Prompting (INTERCPT): Ambiguity Resolution for Crosslingual Conditional Generation with Interaction”, 2023-01-24 ( ; similar)
“U-PaLM: Transcending Scaling Laws With 0.1% Extra Compute”, Et Al 2022
“U-PaLM: Transcending Scaling Laws with 0.1% Extra Compute”, 2022-10-20 ( ; similar; bibliography)
“Challenging BIG-Bench Tasks (BBH) and Whether Chain-of-Thought Can Solve Them”, Et Al 2022
“Challenging BIG-Bench Tasks (BBH) and Whether Chain-of-Thought Can Solve Them”, 2022-10-17 ( ; similar; bibliography)
“Language Models Are Multilingual Chain-of-Thought Reasoners”, Et Al 2022
“Language Models are Multilingual Chain-of-Thought Reasoners”, 2022-10-06 ( ; similar; bibliography)
“Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit”, Et Al 2022
“Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit”, 2022-07-18 ( ; similar)
“Emergent Abilities of Large Language Models”, Et Al 2022
“Emergent Abilities of Large Language Models”, 2022-06-15 (similar)
“Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models”, Et Al 2022
“Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models”, 2022-06-09 ( ; backlinks; similar)
“Data Distributional Properties Drive Emergent Few-Shot Learning in Transformers”, Et Al 2022
“Data Distributional Properties Drive Emergent Few-Shot Learning in Transformers”, 2022-04-22 ( ; similar)
“PaLM: Scaling Language Modeling With Pathways”, Et Al 2022
“PaLM: Scaling Language Modeling with Pathways”, 2022-04-05 ( ; similar; bibliography)
“In-context Learning and Induction Heads”, Et Al 2022
“In-context Learning and Induction Heads”, 2022-03-08 ( )
“Predictability and Surprise in Large Generative Models”, Et Al 2022
“Predictability and Surprise in Large Generative Models”, 2022-02-15 ( ; similar; bibliography)
“The Effects of Reward Misspecification: Mapping and Mitigating Misaligned Models”, Et Al 2022
“The Effects of Reward Misspecification: Mapping and Mitigating Misaligned Models”, 2022-01-10 ( ; backlinks; similar; bibliography)
“A Mathematical Framework for Transformer Circuits”, Et Al 2021
“A Mathematical Framework for Transformer Circuits”, 2021-12-22 ( )
“Scaling Language Models: Methods, Analysis & Insights from Training Gopher”, Et Al 2021
“Scaling Language Models: Methods, Analysis & Insights from Training Gopher”, 2021-12-08 ( ; similar; bibliography)
“A General Language Assistant As a Laboratory for Alignment”, Et Al 2021
“A General Language Assistant as a Laboratory for Alignment”, 2021-12-01 ( ; similar; bibliography)
“Mapping Language Models to Grounded Conceptual Spaces”, 2021
“Mapping Language Models to Grounded Conceptual Spaces”, 2021-11-18 ( ; backlinks; similar; bibliography)
“Program Synthesis With Large Language Models”, Et Al 2021
“Program Synthesis with Large Language Models”, 2021-08-16 ( ; similar)
“MMLU: Measuring Massive Multitask Language Understanding”, Et Al 2020
“MMLU: Measuring Massive Multitask Language Understanding”, 2020-09-07 ( ; backlinks; similar; bibliography)
“GPT-3: Language Models Are Few-Shot Learners”, Et Al 2020
“GPT-3: Language Models are Few-Shot Learners”, 2020-05-28 ( ; similar)
“Emergence in Cognitive Science”, 2010
“Emergence in Cognitive Science”, 2010-09-14 ( ; similar; bibliography)
“The Phase Transition In Human Cognition § Phase Transitions in Language Processing”, Et Al 2009 (page 13)
“The Phase Transition In Human Cognition § Phase Transitions in Language Processing”, 2009-03-01 ( ; similar)
Wikipedia
Miscellaneous
Link Bibliography
-
https://arxiv.org/abs/2210.11399#google
: “U-PaLM: Transcending Scaling Laws With 0.1% Extra Compute”, : -
https://arxiv.org/abs/2210.09261#google
: “Challenging BIG-Bench Tasks (BBH) and Whether Chain-of-Thought Can Solve Them”, : -
https://arxiv.org/abs/2210.03057#google
: “Language Models Are Multilingual Chain-of-Thought Reasoners”, : -
https://arxiv.org/abs/2204.02311#google
: “PaLM: Scaling Language Modeling With Pathways”, : -
https://arxiv.org/abs/2202.07785#anthropic
: “Predictability and Surprise in Large Generative Models”, : -
https://arxiv.org/abs/2201.03544
: “The Effects of Reward Misspecification: Mapping and Mitigating Misaligned Models”, Alexander Pan, Kush Bhatia, Jacob Steinhardt: -
https://arxiv.org/abs/2112.11446#deepmind
: “Scaling Language Models: Methods, Analysis & Insights from Training Gopher”, : -
https://arxiv.org/abs/2112.00861#anthropic
: “A General Language Assistant As a Laboratory for Alignment”, : -
https://openreview.net/forum?id=gJcEM8sxHK
: “Mapping Language Models to Grounded Conceptual Spaces”, Roma Patel, Ellie Pavlick: -
https://arxiv.org/abs/2009.03300
: “MMLU: Measuring Massive Multitask Language Understanding”, Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, Jacob Steinhardt: -
https://onlinelibrary.wiley.com/doi/full/10.1111/j.1756-8765.2010.01116.x
: “Emergence in Cognitive Science”, James L. McClelland: