- See Also
-
Links
- “The Structure of the Token Space for Large Language Models”, Robinson et al 2024
- “OpenAI Begins Training next AI Model As It Battles Safety Concerns: Executive Appears to Backtrack on Start-Up’s Vision of Building ‘Superintelligence’ After Exits from ‘Superalignment’ Team”, Criddle 2024
- “VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?”, Liu et al 2024
- “How Tech Giants Cut Corners to Harvest Data for AI: OpenAI, Google and Meta Ignored Corporate Policies, Altered Their Own Rules and Discussed Skirting Copyright Law As They Sought Online Information to Train Their Newest Artificial Intelligence Systems”, Metz et al 2024
- “Hal-Eval: A Universal and Fine-Grained Hallucination Evaluation Framework for Large Vision Language Models”, Jiang et al 2024
- “Using Hallucinations to Bypass GPT-4’s Filter”, Lemkin 2024
- “The Carbon Emissions of Writing and Illustrating Are Lower for AI Than for Humans”, Tomlinson et al 2024
- “Altman Says ChatGPT Will Have to Evolve in ‘Uncomfortable’ Ways”, Heath 2024
- “RAG vs Fine-Tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture”, Balaguer et al 2024
- “Gemini vs GPT-4-V: A Preliminary Comparison and Combination of Vision-Language Models Through Qualitative Cases”, Qi et al 2023
- “Exploiting Novel GPT-4 APIs”, Pelrine et al 2023
- “ByteDance Is Secretly Using OpenAI’s Tech to Build a Competitor”, Heath 2023
- “Did I Get Sam Altman Fired from OpenAI?: Nathan’s Red-Teaming Experience, Noticing How the Board Was Not Aware of GPT-4 Jailbreaks & Had Not Even Tried GPT-4 prior to Its Early Release”, Labenz 2023
- “Did I Get Sam Altman Fired from OpenAI? § GPT-4-Base”, Labenz 2023
- “Inside the Chaos at OpenAI: Sam Altman’s Weekend of Shock and Drama Began a Year Ago, With the Release of ChatGPT”, Hao & Warzel 2023
- “Instruction-Following Evaluation for Large Language Models”, Zhou et al 2023
- “Everything of Thoughts: Defying the Law of Penrose Triangle for Thought Generation”, Ding et al 2023
- “Holistic Analysis of Hallucination in GPT-4-V(ision): Bias and Interference Challenges”, Cui et al 2023
- “Scalable and Transferable Black-Box Jailbreaks for Language Models via Persona Modulation”, Shah et al 2023
- “Does GPT-4 Pass the Turing Test?”, Jones & Bergen 2023
- “The Generative AI Paradox: "What It Can Create, It May Not Understand"”, West et al 2023
- “Interpreting Learned Feedback Patterns in Large Language Models”, Marks et al 2023
- “PAIR: Jailbreaking Black Box Large Language Models in 20 Queries”, Chao et al 2023
- “The Dawn of LMMs: Preliminary Explorations With GPT-4-V(ision)”, Yang et al 2023
- “Summarization Is (Almost) Dead”, Pu et al 2023
- “Non-Determinism in GPT-4 Is Caused by Sparse MoE”, 152334H 2023
- “Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations”, Chen et al 2023
- “Large Language Models As Superpositions of Cultural Perspectives”, Kovač et al 2023
- “Large Language Models Understand and Can Be Enhanced by Emotional Stimuli”, Li et al 2023
- “AI Is a Lot of Work: As the Technology Becomes Ubiquitous, a Vast Tasker Underclass Is Emerging—And Not Going Anywhere”, Dzieza 2023
- “Why YouTube Could Give Google an Edge in AI”, Victor 2023
- “LTX by Broadridge Launches BondGPT™ Powered by OpenAI GPT-4”, Inc 2023
- “LIMA: Less Is More for Alignment”, Zhou et al 2023
- “Tree of Thoughts (ToT): Deliberate Problem Solving With Large Language Models”, Yao et al 2023
- “Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4”, Chang et al 2023
- “What’s AGI, and Why Are AI Experts Skeptical? ChatGPT and Other Bots Have Revived Conversations on Artificial General Intelligence. Scientists Say Algorithms Won’t Surpass You Any Time Soon”, Rogers 2023
- “ChemCrow: Augmenting Large-Language Models With Chemistry Tools”, Bran et al 2023
- “Introducing Microsoft 365 Copilot—Your Copilot for Work”, Spataro 2023
- “Sam Altman on What Makes Him ‘Super Nervous’ About AI: The OpenAI Co-Founder Thinks Tools like GPT-4 Will Be Revolutionary. But He’s Wary of Downsides”, Swisher 2023
- “GPT-4 Technical Report”, OpenAI 2023
- “GPT-4 Is Coming next Week—And It Will Be Multimodal, Says Microsoft Germany”, Hahn 2023
- “As AI Booms, Lawmakers Struggle to Understand the Technology: Tech Innovations Are Again Racing ahead of Washington’s Ability to Regulate Them, Lawmakers and AI Experts Said”, Kang & Satariano 2023
- “Allen & Overy Breaks the Internet (and New Ground) With Co-Pilot Harvey”, Hill 2023
- davidtayar5 @ "2023-02-10"
- “New Bing, and an Interview With Kevin Scott and Sam Altman About the Microsoft-OpenAI Partnership”, Thompson 2023
- “How ChatGPT Kicked Off an AI Arms Race: Even inside the Company, the Chatbot’s Popularity Has Come As Something of a Shock”, Roose 2023
- “Connor Leahy on Aliens, Ethics, Economics, Memetics, and Education § GPT-4”, Leahy 2023
- “OpenAI CEO Sam Altman on GPT-4: ‘People Are Begging to Be Disappointed and They Will Be’”, Vincent 2023
- “Microsoft Bets Big on the Creator of ChatGPT in Race to Dominate AI: As a New Chatbot Wows the World With Its Conversational Talents, a Resurgent Tech Giant Is Poised to Reap the Benefits While Doubling down on a Relationship With the Start-Up OpenAI”, Metz & Weise 2023
- “What to Expect When You’re Expecting…GPT-4. What Comes After ChatGPT? 7 Predictions for 2023 § GPT-4”, Marcus 2022
- “Here’s What I Saw at an AI Hackathon: AI Gossip, Celebrity Sightings, Tech Trends—And Some Great Projects”, Shipper 2022
- “6 New Theories About AI: Software With Superpowers § GPT-4”, Armstrong 2022
- immad @ "2022-11-22"
- “GPT-4 Rumors From Silicon Valley: People Are Saying Things…”, Romero 2022
- EMostaque @ "2022-08-27"
- “Chinchilla: Training Compute-Optimal Large Language Models”, Hoffmann et al 2022
- “Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer”, Yang et al 2022
- “Sparse Is Enough in Scaling Transformers”, Jaszczur et al 2021
- “An Analysis of AI Political Preferences from a European Perspective”
- Sort By Magic
- Miscellaneous
- Bibliography
See Also
Links
“The Structure of the Token Space for Large Language Models”, Robinson et al 2024
“OpenAI Begins Training next AI Model As It Battles Safety Concerns: Executive Appears to Backtrack on Start-Up’s Vision of Building ‘Superintelligence’ After Exits from ‘Superalignment’ Team”, Criddle 2024
“VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?”, Liu et al 2024
VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?
“How Tech Giants Cut Corners to Harvest Data for AI: OpenAI, Google and Meta Ignored Corporate Policies, Altered Their Own Rules and Discussed Skirting Copyright Law As They Sought Online Information to Train Their Newest Artificial Intelligence Systems”, Metz et al 2024
“Hal-Eval: A Universal and Fine-Grained Hallucination Evaluation Framework for Large Vision Language Models”, Jiang et al 2024
“Using Hallucinations to Bypass GPT-4’s Filter”, Lemkin 2024
“The Carbon Emissions of Writing and Illustrating Are Lower for AI Than for Humans”, Tomlinson et al 2024
The carbon emissions of writing and illustrating are lower for AI than for humans
“Altman Says ChatGPT Will Have to Evolve in ‘Uncomfortable’ Ways”, Heath 2024
Altman says ChatGPT will have to evolve in ‘uncomfortable’ ways
“RAG vs Fine-Tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture”, Balaguer et al 2024
RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture
“Gemini vs GPT-4-V: A Preliminary Comparison and Combination of Vision-Language Models Through Qualitative Cases”, Qi et al 2023
“Exploiting Novel GPT-4 APIs”, Pelrine et al 2023
“ByteDance Is Secretly Using OpenAI’s Tech to Build a Competitor”, Heath 2023
ByteDance is secretly using OpenAI’s tech to build a competitor
“Did I Get Sam Altman Fired from OpenAI?: Nathan’s Red-Teaming Experience, Noticing How the Board Was Not Aware of GPT-4 Jailbreaks & Had Not Even Tried GPT-4 prior to Its Early Release”, Labenz 2023
“Did I Get Sam Altman Fired from OpenAI? § GPT-4-Base”, Labenz 2023
“Inside the Chaos at OpenAI: Sam Altman’s Weekend of Shock and Drama Began a Year Ago, With the Release of ChatGPT”, Hao & Warzel 2023
“Instruction-Following Evaluation for Large Language Models”, Zhou et al 2023
“Everything of Thoughts: Defying the Law of Penrose Triangle for Thought Generation”, Ding et al 2023
Everything of Thoughts: Defying the Law of Penrose Triangle for Thought Generation
“Holistic Analysis of Hallucination in GPT-4-V(ision): Bias and Interference Challenges”, Cui et al 2023
Holistic Analysis of Hallucination in GPT-4-V(ision): Bias and Interference Challenges
“Scalable and Transferable Black-Box Jailbreaks for Language Models via Persona Modulation”, Shah et al 2023
Scalable and Transferable Black-Box Jailbreaks for Language Models via Persona Modulation
“Does GPT-4 Pass the Turing Test?”, Jones & Bergen 2023
“The Generative AI Paradox: "What It Can Create, It May Not Understand"”, West et al 2023
The Generative AI Paradox: "What It Can Create, It May Not Understand"
“Interpreting Learned Feedback Patterns in Large Language Models”, Marks et al 2023
Interpreting Learned Feedback Patterns in Large Language Models
“PAIR: Jailbreaking Black Box Large Language Models in 20 Queries”, Chao et al 2023
PAIR: Jailbreaking Black Box Large Language Models in 20 Queries
“The Dawn of LMMs: Preliminary Explorations With GPT-4-V(ision)”, Yang et al 2023
The Dawn of LMMs: Preliminary Explorations with GPT-4-V(ision)
“Summarization Is (Almost) Dead”, Pu et al 2023
“Non-Determinism in GPT-4 Is Caused by Sparse MoE”, 152334H 2023
“Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations”, Chen et al 2023
Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations
“Large Language Models As Superpositions of Cultural Perspectives”, Kovač et al 2023
Large Language Models as Superpositions of Cultural Perspectives
“Large Language Models Understand and Can Be Enhanced by Emotional Stimuli”, Li et al 2023
Large Language Models Understand and Can be Enhanced by Emotional Stimuli
“AI Is a Lot of Work: As the Technology Becomes Ubiquitous, a Vast Tasker Underclass Is Emerging—And Not Going Anywhere”, Dzieza 2023
“Why YouTube Could Give Google an Edge in AI”, Victor 2023
“LTX by Broadridge Launches BondGPT™ Powered by OpenAI GPT-4”, Inc 2023
“LIMA: Less Is More for Alignment”, Zhou et al 2023
“Tree of Thoughts (ToT): Deliberate Problem Solving With Large Language Models”, Yao et al 2023
Tree of Thoughts (ToT): Deliberate Problem Solving with Large Language Models
“Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4”, Chang et al 2023
Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4
“What’s AGI, and Why Are AI Experts Skeptical? ChatGPT and Other Bots Have Revived Conversations on Artificial General Intelligence. Scientists Say Algorithms Won’t Surpass You Any Time Soon”, Rogers 2023
“ChemCrow: Augmenting Large-Language Models With Chemistry Tools”, Bran et al 2023
ChemCrow: Augmenting large-language models with chemistry tools
“Introducing Microsoft 365 Copilot—Your Copilot for Work”, Spataro 2023
“Sam Altman on What Makes Him ‘Super Nervous’ About AI: The OpenAI Co-Founder Thinks Tools like GPT-4 Will Be Revolutionary. But He’s Wary of Downsides”, Swisher 2023
“GPT-4 Technical Report”, OpenAI 2023
“GPT-4 Is Coming next Week—And It Will Be Multimodal, Says Microsoft Germany”, Hahn 2023
GPT-4 is coming next week—and it will be multimodal, says Microsoft Germany
“As AI Booms, Lawmakers Struggle to Understand the Technology: Tech Innovations Are Again Racing ahead of Washington’s Ability to Regulate Them, Lawmakers and AI Experts Said”, Kang & Satariano 2023
“Allen & Overy Breaks the Internet (and New Ground) With Co-Pilot Harvey”, Hill 2023
Allen & Overy breaks the internet (and new ground) with co-pilot Harvey
davidtayar5 @ "2023-02-10"
Context on the NVIDIA ChatGPT opportunity—and ramifications of large language model enthusiasm
“New Bing, and an Interview With Kevin Scott and Sam Altman About the Microsoft-OpenAI Partnership”, Thompson 2023
New Bing, and an Interview with Kevin Scott and Sam Altman About the Microsoft-OpenAI Partnership
“How ChatGPT Kicked Off an AI Arms Race: Even inside the Company, the Chatbot’s Popularity Has Come As Something of a Shock”, Roose 2023
“Connor Leahy on Aliens, Ethics, Economics, Memetics, and Education § GPT-4”, Leahy 2023
Connor Leahy on Aliens, Ethics, Economics, Memetics, and Education § GPT-4
“OpenAI CEO Sam Altman on GPT-4: ‘People Are Begging to Be Disappointed and They Will Be’”, Vincent 2023
OpenAI CEO Sam Altman on GPT-4: ‘people are begging to be disappointed and they will be’
“Microsoft Bets Big on the Creator of ChatGPT in Race to Dominate AI: As a New Chatbot Wows the World With Its Conversational Talents, a Resurgent Tech Giant Is Poised to Reap the Benefits While Doubling down on a Relationship With the Start-Up OpenAI”, Metz & Weise 2023
“What to Expect When You’re Expecting…GPT-4. What Comes After ChatGPT? 7 Predictions for 2023 § GPT-4”, Marcus 2022
What to Expect When You’re Expecting…GPT-4. What comes after ChatGPT? 7 predictions for 2023 § GPT-4
“Here’s What I Saw at an AI Hackathon: AI Gossip, Celebrity Sightings, Tech Trends—And Some Great Projects”, Shipper 2022
“6 New Theories About AI: Software With Superpowers § GPT-4”, Armstrong 2022
immad @ "2022-11-22"
Have talked to several people who have used GPT-4, they all say it is uncanny how good it is
“GPT-4 Rumors From Silicon Valley: People Are Saying Things…”, Romero 2022
EMostaque @ "2022-08-27"
Stack more layers is fine as GPT-4 is about to show but there are superior routes…
“Chinchilla: Training Compute-Optimal Large Language Models”, Hoffmann et al 2022
“Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer”, Yang et al 2022
Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
“Sparse Is Enough in Scaling Transformers”, Jaszczur et al 2021
“An Analysis of AI Political Preferences from a European Perspective”
An Analysis of AI Political Preferences from a European Perspective
Sort By Magic
Annotations sorted by machine learning into inferred 'tags'. This provides an alternative way to browse: instead of by date order, one can browse in topic order. The 'sorted' list has been automatically clustered into multiple sections & auto-labeled for easier browsing.
Beginning with the newest annotation, it uses the embedding of each annotation to attempt to create a list of nearest-neighbor annotations, creating a progression of topics. For more details, see the link.
ai-ethics
instruction-following
multimodal
hallucination-issues
gpt-future
Miscellaneous
-
/doc/ai/nn/transformer/gpt/4/2024-06-30-openai-chatgpt4o-systemprompt.txt
: -
https://github.com/spdustin/ChatGPT-AutoExpert/blob/main/System%20Prompts.md
-
https://newsletter.pragmaticengineer.com/p/scaling-chatgpt#%C2%A7five-scaling-challenges
: -
https://nostalgebraist.tumblr.com/post/705192637617127424/gpt-4-prediction-it-wont-be-very-useful
: -
https://openai.com/blog/new-models-and-developer-products-announced-at-devday
: -
https://simonwillison.net/2023/Oct/14/multi-modal-prompt-injection/
-
https://terrytao.wordpress.com/2023/06/19/ai-anthology/#comment-678803
: -
https://www.lesswrong.com/posts/DaaFce3hBoEzYhdvz/how-well-did-manifold-predict-gpt-4
: -
https://www.lesswrong.com/posts/qdStMFDMrWAnTqNWL/gpt-4-predictions
:View External Link:
https://www.lesswrong.com/posts/qdStMFDMrWAnTqNWL/gpt-4-predictions
-
https://www.lesswrong.com/posts/sY3a4Rfa48CgteBEm/chatgpt-can-learn-indirect-control
: -
https://www.lesswrong.com/posts/t9svvNPNmFf5Qa3TA/mysteries-of-mode-collapse#pfHTedu4GKaWoxD5K
-
https://www.lesswrong.com/posts/tbJdxJMAiehewGpq2/impressions-from-base-gpt-4
-
https://www.lesswrong.com/posts/tt7WtqiEyEiLmAecZ/what-will-gpt-4-be-incapable-of
: -
https://www.linkedin.com/pulse/building-new-bing-jordi-ribas/
: -
https://www.metaculus.com/questions/14305/when-will-gpt-4-be-announced/
-
https://www.reddit.com/r/singularity/comments/1dmap17/possible_timelines_for_gpt45_and_gpt5/
: -
https://www.wired.com/story/cerebras-chip-cluster-neural-networks-ai/
:
Bibliography
-
https://arxiv.org/abs/2410.08993
: “The Structure of the Token Space for Large Language Models”, -
https://arxiv.org/abs/2404.05955
: “VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?”, -
https://www.nytimes.com/2024/04/06/technology/tech-giants-harvest-data-artificial-intelligence.html
: “How Tech Giants Cut Corners to Harvest Data for AI: OpenAI, Google and Meta Ignored Corporate Policies, Altered Their Own Rules and Discussed Skirting Copyright Law As They Sought Online Information to Train Their Newest Artificial Intelligence Systems”, -
https://www.nature.com/articles/s41598-024-54271-x
: “The Carbon Emissions of Writing and Illustrating Are Lower for AI Than for Humans”, -
https://arxiv.org/abs/2401.08406#microsoft
: “RAG vs Fine-Tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture”, -
https://arxiv.org/abs/2312.15011
: “Gemini vs GPT-4-V: A Preliminary Comparison and Combination of Vision-Language Models Through Qualitative Cases”, -
https://cognitiverevolution.substack.com/p/did-i-get-sam-altman-fired-from-openai
: “Did I Get Sam Altman Fired from OpenAI?: Nathan’s Red-Teaming Experience, Noticing How the Board Was Not Aware of GPT-4 Jailbreaks & Had Not Even Tried GPT-4 prior to Its Early Release”, -
https://www.theatlantic.com/technology/archive/2023/11/sam-altman-open-ai-chatgpt-chaos/676050/
: “Inside the Chaos at OpenAI: Sam Altman’s Weekend of Shock and Drama Began a Year Ago, With the Release of ChatGPT”, -
https://arxiv.org/abs/2311.03287
: “Holistic Analysis of Hallucination in GPT-4-V(ision): Bias and Interference Challenges”, -
https://arxiv.org/abs/2310.08419
: “PAIR: Jailbreaking Black Box Large Language Models in 20 Queries”, -
https://arxiv.org/abs/2309.17421
: “The Dawn of LMMs: Preliminary Explorations With GPT-4-V(ision)”, -
https://152334h.github.io/blog/non-determinism-in-gpt-4/
: “Non-Determinism in GPT-4 Is Caused by Sparse MoE”, -
https://arxiv.org/abs/2307.07870
: “Large Language Models As Superpositions of Cultural Perspectives”, -
https://arxiv.org/abs/2307.11760#microsoft
: “Large Language Models Understand and Can Be Enhanced by Emotional Stimuli”, -
https://www.theverge.com/features/23764584/ai-artificial-intelligence-data-notation-labor-scale-surge-remotasks-openai-chatbots
: “AI Is a Lot of Work: As the Technology Becomes Ubiquitous, a Vast Tasker Underclass Is Emerging—And Not Going Anywhere”, -
https://www.theinformation.com/articles/why-youtube-could-give-google-an-edge-in-ai
: “Why YouTube Could Give Google an Edge in AI”, -
https://arxiv.org/abs/2305.10601#deepmind
: “Tree of Thoughts (ToT): Deliberate Problem Solving With Large Language Models”, -
https://www.wired.com/story/what-is-artificial-general-intelligence-agi-explained/
: “What’s AGI, and Why Are AI Experts Skeptical? ChatGPT and Other Bots Have Revived Conversations on Artificial General Intelligence. Scientists Say Algorithms Won’t Surpass You Any Time Soon”, -
https://blogs.microsoft.com/blog/2023/03/16/introducing-microsoft-365-copilot-your-copilot-for-work/
: “Introducing Microsoft 365 Copilot—Your Copilot for Work”, -
https://nymag.com/intelligencer/2023/03/on-with-kara-swisher-sam-altman-on-the-ai-revolution.html
: “Sam Altman on What Makes Him ‘Super Nervous’ About AI: The OpenAI Co-Founder Thinks Tools like GPT-4 Will Be Revolutionary. But He’s Wary of Downsides”, -
https://www.heise.de/news/GPT-4-is-coming-next-week-and-it-will-be-multimodal-says-Microsoft-Germany-7540972.html
: “GPT-4 Is Coming next Week—And It Will Be Multimodal, Says Microsoft Germany”, -
https://www.nytimes.com/2023/03/03/technology/artificial-intelligence-regulation-congress.html
: “As AI Booms, Lawmakers Struggle to Understand the Technology: Tech Innovations Are Again Racing ahead of Washington’s Ability to Regulate Them, Lawmakers and AI Experts Said”, -
https://legaltechnology.com/2023/02/16/allen-overy-breaks-the-internet-and-new-ground-with-co-pilot-harvey/
: “Allen & Overy Breaks the Internet (and New Ground) With Co-Pilot Harvey”, -
https://x.com/davidtayar5/status/1627690520456691712
: “Context on the NVIDIA ChatGPT Opportunity—And Ramifications of Large Language Model Enthusiasm”, -
https://stratechery.com/2023/new-bing-and-an-interview-with-kevin-scott-and-sam-altman-about-the-microsoft-openai-partnership/
: “New Bing, and an Interview With Kevin Scott and Sam Altman About the Microsoft-OpenAI Partnership”, -
https://www.nytimes.com/2023/02/03/technology/chatgpt-openai-artificial-intelligence.html
: “How ChatGPT Kicked Off an AI Arms Race: Even inside the Company, the Chatbot’s Popularity Has Come As Something of a Shock”, -
https://www.theverge.com/23560328/openai-gpt-4-rumor-release-date-sam-altman-interview
: “OpenAI CEO Sam Altman on GPT-4: ‘People Are Begging to Be Disappointed and They Will Be’”, -
https://www.nytimes.com/2023/01/12/technology/microsoft-openai-chatgpt.html
: “Microsoft Bets Big on the Creator of ChatGPT in Race to Dominate AI: As a New Chatbot Wows the World With Its Conversational Talents, a Resurgent Tech Giant Is Poised to Reap the Benefits While Doubling down on a Relationship With the Start-Up OpenAI”, -
https://garymarcus.substack.com/p/what-to-expect-when-youre-expecting
: “What to Expect When You’re Expecting…GPT-4. What Comes After ChatGPT? 7 Predictions for 2023 § GPT-4”, -
https://every.to/chain-of-thought/the-knee-of-the-exponential-curve
: “Here’s What I Saw at an AI Hackathon: AI Gossip, Celebrity Sightings, Tech Trends—And Some Great Projects”, -
https://www.thealgorithmicbridge.com/p/gpt-4-rumors-from-silicon-valley
: “GPT-4 Rumors From Silicon Valley: People Are Saying Things…”, -
https://x.com/EMostaque/status/1563715319566409729
: “Stack More Layers Is Fine As GPT-4 Is about to Show but There Are Superior Routes…”, -
https://arxiv.org/abs/2203.15556#deepmind
: “Chinchilla: Training Compute-Optimal Large Language Models”, -
https://arxiv.org/abs/2203.03466#microsoft
: “Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer”, -
https://arxiv.org/abs/2111.12763#google
: “Sparse Is Enough in Scaling Transformers”,