Bibliography:

  1. ‘GPT’ tag

  2. ‘GPT-3 fiction’ tag

  3. ‘GPT-3 humor’ tag

  4. ‘GPT-3 nonfiction’ tag

  5. ‘GPT-3 poetry’ tag

  6. The Scaling Hypothesis

  7. Benchmarking the Performance of Large Language Models on the Cerebras Wafer Scale Engine

  8. RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture

  9. Inside the Chaos at OpenAI: Sam Altman’s weekend of shock and drama began a year ago, with the release of ChatGPT

  10. Everything of Thoughts: Defying the Law of Penrose Triangle for Thought Generation

  11. Does GPT-4 Pass the Turing Test?

  12. PAIR: Jailbreaking Black Box Large Language Models in 20 Queries

  13. Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!

  14. Non-determinism in GPT-4 is caused by Sparse MoE

  15. Large Language Models as Superpositions of Cultural Perspectives

  16. AI Is a Lot of Work: As the technology becomes ubiquitous, a vast tasker underclass is emerging—and not going anywhere

  17. I’m Afraid I Can’t Do That: Predicting Prompt Refusal in Black-Box Generative Language Models

  18. Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4

  19. GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models

  20. Why didn’t DeepMind build GPT-3?

  21. OpenAI’s Sam Altman Talks ChatGPT And How Artificial General Intelligence Can ‘Break Capitalism’

  22. GPT-3 as Knowledge Worker: A Zero-Shot Evaluation of AI CPA Capabilities

  23. Language models are better than humans at next-token prediction

  24. HALIE: Evaluating Human-Language Model Interaction

  25. TruthfulQA: Measuring How Models Mimic Human Falsehoods

  26. ‘How GPT-3 Is Shaping Our AI Future’ With Sam Altman/Azeem Azhar (The Exponential View), Wednesday 7 October 2020

  27. Scaling Laws for Neural Language Models: Figure 15: Far beyond the Model Sizes We Study Empirically, We Find a Contradiction between Our Equations § Pg17

  28. 20d126b9c3baf640f8d1d5dff3e253faac2e8242.pdf#page=17&org=openai

  29. Towards Synthesizing Complex Programs from Input-Output Examples

  30. Genetics of caffeine consumption and responses to caffeine

  31. Why GPT-3 Matters

  32. Greg Brockman: OpenAI and AGI

  33. GPT-3 Gives Some Interesting True and False Answers to Some Questions. But It’s Important to Note That It Gives opposite Answers Just As Often, I Cheery Picked the Most ‘Sensational’ Ones. Usually It Said the opposite Thing, and It Also Role-Plays Sometimes (eg. As a Spy)

  34. I Gave GPT-3 Access to Chrome With the Objective ‘Please Buy Me AirPods’...It Successfully Made It to the Product Page, but Got Sidetracked With Walmart’s Privacy Policy. Since Even a Simplified DOM Is Far Too Large for a Single Prompt, Multiple Prompts Are given Different Chunks of the DOM, Each Generating Their Own ‘Interaction’. Another Prompt Then Takes All the Proposed Interactions and Selects the Best One, Sort of like a Tournament Bracket. For More Complex Web Pages, the Time It Takes to Generate an Action Scales at 𝒪(log n) With the Size of the DOM—Really Fast! It Also Gets around Token Limits, so You Could Technically Process an Infinitely Large DOM!

  35. 2019-11-07-amodei-aiandcompute-twodistincteras-gpt3modified.jpg

  36. https://andrewmayne.com/2023/11/14/is-the-reversal-curse-real/

  37. https://barryzhang.substack.com/p/our-humble-attempt-at-fine-tuning

  38. https://openai.com/blog/gpt-3-5-turbo-fine-tuning-and-api-updates

  39. https://openai.com/pricing#fine-tuning-models

  40. https://osf.io/preprints/psyarxiv/5b26t

  41. 89baa2c3bb8bf25c08aa4f0505c7face30a79917.html

  42. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4594466

  43. https://www.cerebras.net/blog/introducing-gigagpt-gpt-3-sized-models-in-565-lines-of-code

  44. dfda8a66e3496fa3145e60fc5d38aaee429283be.html

  45. https://www.haihai.ai/pen15/

  46. https://www.lesswrong.com/posts/CNPvESPru3XNqsw7A/what-s-up-with-all-the-non-mormons-weirdly-specific

  47. https://www.lesswrong.com/posts/c6uTNm5erRrmyJvvD/mapping-the-semantic-void-strange-goings-on-in-gpt-embedding

  48. https://www.lesswrong.com/posts/t9svvNPNmFf5Qa3TA/mysteries-of-mode-collapse#pfHTedu4GKaWoxD5K

  49. https://www.lesswrong.com/posts/tJAD2LG9uweeEfjwq/estimating-efficiency-improvements-in-llm-pre-training

  50. https://www.reddit.com/r/mlscaling/comments/146rgq2/chatgpt_is_running_quantized/

  51. d8955b8f7f77c41587434170074e66cee41cf31c.html

  52. https://x.com/DanNeidle/status/1664613427472375808

  53. https://x.com/_akhaliq/status/1667030273270087681

  54. https://x.com/somefoundersalt/status/1708599134960398586

  55. https://x.com/yoheinakajima/status/1670557048743010305

  56. RAG vs Fine-tuning: Pipelines, Tradeoffs, and a Case Study on Agriculture

  57. https%253A%252F%252Farxiv.org%252Fabs%252F2401.08406%2523microsoft.html

  58. Inside the Chaos at OpenAI: Sam Altman’s weekend of shock and drama began a year ago, with the release of ChatGPT

  59. https%253A%252F%252Fwww.theatlantic.com%252Ftechnology%252Farchive%252F2023%252F11%252Fsam-altman-open-ai-chatgpt-chaos%252F676050%252F.html

  60. PAIR: Jailbreaking Black Box Large Language Models in 20 Queries

  61. https%253A%252F%252Farxiv.org%252Fabs%252F2310.08419.html

  62. Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!

  63. https%253A%252F%252Farxiv.org%252Fabs%252F2310.03693.html

  64. Non-determinism in GPT-4 is caused by Sparse MoE

  65. https%253A%252F%252F152334h.github.io%252Fblog%252Fnon-determinism-in-gpt-4%252F.html

  66. Large Language Models as Superpositions of Cultural Perspectives

  67. https%253A%252F%252Farxiv.org%252Fabs%252F2307.07870.html

  68. AI Is a Lot of Work: As the technology becomes ubiquitous, a vast tasker underclass is emerging—and not going anywhere

  69. https%253A%252F%252Fwww.theverge.com%252Ffeatures%252F23764584%252Fai-artificial-intelligence-data-notation-labor-scale-surge-remotasks-openai-chatbots.html

  70. I’m Afraid I Can’t Do That: Predicting Prompt Refusal in Black-Box Generative Language Models

  71. https%253A%252F%252Farxiv.org%252Fabs%252F2306.03423.html

  72. GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models

  73. https://www.danielianrock.com/

  74. https%253A%252F%252Farxiv.org%252Fabs%252F2303.10130.html

  75. OpenAI’s Sam Altman Talks ChatGPT And How Artificial General Intelligence Can ‘Break Capitalism’

  76. https%253A%252F%252Fwww.forbes.com%252Fsites%252Falexkonrad%252F2023%252F02%252F03%252Fexclusive-openai-sam-altman-chatgpt-agi-google-search%252F.html

  77. GPT-3 as Knowledge Worker: A Zero-Shot Evaluation of AI CPA Capabilities

  78. https%253A%252F%252Farxiv.org%252Fabs%252F2301.04408.html

  79. TruthfulQA: Measuring How Models Mimic Human Falsehoods

  80. Jacob Hilton's Homepage

  81. Owain Evans, AI Alignment Researcher

  82. https%253A%252F%252Farxiv.org%252Fabs%252F2109.07958.html