-
RNN Metadata for Mimicking Author Style
-
Poems
-
GPT-3 Creative Fiction
-
twdne#text
[Transclude the forward-link's context]
-
A Very Unlikely Chess Game
-
Update: Upgrading to 1.5B GPT-2, and adding 22 new subreddit-bots
-
GPT-3: Language Models are Few-Shot Learners
-
Better Language Models and Their Implications
-
https://research.google/blog/transformer-a-novel-neural-network-architecture-for-language-understanding/
-
The Illustrated Transformer
-
The Illustrated GPT-2 (Visualizing Transformer Language Models)
-
The Transformer—Attention Is All You Need.
-
https://blog.floydhub.com/the-transformer-in-pytorch/
-
https://e2eml.school/transformers.html
-
Attention Is All You Need
-
The Annotated Transformer
-
Self-Attention with Relative Position Representations
-
Character-Level Language Modeling with Deeper Self-Attention
-
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
-
Transformer-XL—Combining Transformers and RNNs Into a State-Of-The-Art Language Model
-
Understanding BERT Transformer: Attention Isn’t All You Need
-
Transformers are a very exciting family of machine learning architectures
-
https://amaarora.github.io/2020/02/18/annotatedGPT2.html
-
The Transformer Family: Attention and Self-Attention · Multi-Head Self-Attention · Transformer · Adaptive Computation Time (ACT) · Improved Attention Span: (Longer Attention Span (Transformer-XL) / Adaptive Attention Span / Localized Attention Span (Image Transformer)) · Less Time and Memory Cost: (Sparse Attention Matrix Factorization (Sparse Transformers) / Locality-Sensitive Hashing (Reformer)) · Make It Recurrent (Universal Transformer) · Stabilization for RL (GTrXL)
-
The Bottom-up Evolution of Representations in the Transformer: A Study with Machine Translation and Language Modeling Objectives
-
Karpathy/minGPT: A Minimal PyTorch Re-Implementation of the OpenAI GPT (Generative Pretrained Transformer) Training
-
RASP: Thinking Like Transformers
-
Efficient Attention: Breaking The Quadratic Transformer Bottleneck
-
‘MLP NN’ directory
-
GPT-1: Improving Language Understanding with Unsupervised Learning
-
Language Modeling State-of-the-art leaderboards
-
Language Models are Unsupervised Multitask Learners
-
Humans Who Are Not Concentrating Are Not General Intelligences
-
Gpt-2-Samples
-
LM Explorer (alpha)
-
GPT-2: 6-Month Follow-Up
-
GPT-2: 1.5B Release
-
OpenGPT-2: We Replicated GPT-2-1.5b Because You Can Too
-
https://colab.research.google.com/drive/1BXry0kcm869-RVHHiY6NZmY9uBzbkf1Q
-
GROVER: Defending Against Neural Fake News
-
XLNet: Generalized Autoregressive Pretraining for Language Understanding
-
MegatronLM: Training Billion+ Parameter Language Models Using GPU Model Parallelism
-
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
-
https://colab.research.google.com/drive/1-ROO7L09EupLFLQM-TWgDHa5-FIOdLLh
-
DialoGPT: Large-Scale Generative Pre-training for Conversational Response Generation
-
This Waifu Does Not Exist
-
Talking to Myself or How I Trained GPT-2-1.5b for Rubber Ducking Using My Facebook Chat Data: Using Only Google Colab
-
https://www.thisstorydoesnotexist.com/
-
https://web.archive.org/web/20200110223938/https://stackroboflow.com/
-
Howl
-
An Eternal Howl
-
https://www.reddit.com/r/slatestarcodex/comments/as8ke7/an_eternal_howl/
-
GPT-2 Howl
-
GPT-2 Writes a Shelley Poem
-
GPT-2 As Step Toward General Intelligence
-
First line of famous poems continued by GPT-2
-
gpt-2-poetry
-
Ask GPT-2
-
Ask GPT-2
-
FridAI: ‘Water, water, everywhere’, as read by Artificial Intelligence
-
The Poetry Machine
-
GPT-based Generation for Classical Chinese Poetry
-
Three More GPT-2 Poems
-
https://www.reddit.com/r/MachineLearning/comments/coc09l/p_these_lyrics_do_not_exist/
-
Testing The Limits of GROVER The Neural Fake News Detector. Can It Write Fiction? Can It Write Riddles?
-
https://www.reddit.com/r/SubSimulatorGPT2Meta/comments/ccvspt/update_experimenting_with_generating_hybrid/
-
CTRL: A Conditional Transformer Language Model For Controllable Generation
-
Conditional Transformer Language Model for Controllable Generation
-
https://papergains.co/pdfs/Transformer_Poetry-978-1-7341647-0-1.pdf#page=3
-
345M-GPT-2 After James Wright: Can AI Generate Convincing Contemporary Poetry?
-
GPT-2 AI Poetry Generation: Writing like Donne
-
Writing the Next American Hit: Using GPT-2 to Explore the Possibility of Creating Successful AI-Generated Song Lyrics Possibility of Creating Successful AI-Generated Song Lyric
-
How to Train It
-
Nshepperd/gpt-2: Code for the Paper "Language Models Are Unsupervised Multitask Learners"
-
ConnorJL/GPT2: An Implementation of Training for GPT-2, Supports TPUs
-
Replicating GPT-2-1.5B
-
Addendum: Evaluation of My Model
-
A Corpus of Poetry from Project Gutenberg
-
Dataset Search
-
Poems from Poetryfoundation.org
-
A Small Module Meant for Use in Text Generators That Lets You Filter Strings for Bad Words
-
Success
-
Unhandled Arguments Checked After Execution, Not Before
-
The Curious Case of Neural Text Degeneration
-
https://www.trentonbricken.com/Tail-Free-Sampling/
-
The Unreasonable Effectiveness of Recurrent Neural Networks
-
2019-03-06-gwern-gpt2-poetry-projectgutenberg-network-519407.tar.xz
-
2019-03-06-gpt2-poetry-1000samples.txt
-
https://x.com/theshawwn
-
Kaggle: Your Home for Data Science
-
rnn-metadata#inline-metadata-trick
[Transclude the forward-link's context]
-
2019-10-18-Poetryfoundation-Formatted.txt
-
2019-10-17-117m-poetry-cleanprojectgutenberg-samples.txt
-
2019-10-19-117m-poetryfoundation-samples.txt
-
2019-10-19-gwern-gpt2-poetry-pgclean-117m.tar.xz
-
2019-03-06-gwern-gpt2-poetry-prefix-projectgutenberg-network-224474.tar.xz
-
2019-03-06-gpt2-poetry-prefix-1000samples.txt
-
To a Skylark by Percy Bysshe Shelley
-
Gwern’s AI-Generated Poetry
-
Overview for Starspawn0
-
Semantic projection: recovering human knowledge of multiple, distinct object features from word embeddings
-
Distributional Vectors Encode Referential Attributes
-
Dynamic Word Embeddings for Evolving Semantic Discovery
-
Verb Physics: Relative Physical Knowledge of Actions and Objects
-
Language Models Represent Space and Time
-
Language Encodes Geographical Information
-
Grounding the Ungrounded: Estimating Locations of Unknown Place Names from Linguistic Associations and Grounded Representations
-
Books by Pope, Alexander (Sorted by Popularity)
-
2019-03-16-gpt2-poetry-prefix-jabberwocky-100samples.txt
-
The Jingle Book by Carolyn Wells
-
https://openai.com/index/better-language-models/#update
-
UniLM: Unified Language Model Pre-training for Natural Language Understanding and Generation
-
Fitting Larger Networks into Memory: TLDR; We Release the Python/Tensorflow Package Openai/gradient-Checkpointing, That Lets You Fit 10× Larger Neural Nets into Memory at the Cost of an Additional 20% Computation Time
-
Generating Long Sequences with Sparse Transformers
-
Training Deep Nets with Sublinear Memory Cost
-
Memory-Efficient Backpropagation through Time
-
MuseNet: a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country to Mozart to the Beatles
-
Why Momentum Really Works
-
Cyclical Learning Rates for Training Neural Networks
-
SGDR: Stochastic Gradient Descent with Warm Restarts
-
Averaging Weights Leads to Wider Optima and Better Generalization
-
2019-05-13-gwern-gpt2-poetry-345m.tar.xz
-
2019-05-13-gpt2-poetry-345m-5000samples.txt
-
Reassuring
-
This Is a Python Script As Described in XKCD #1263: ‘Reassuring’. It Generates Thousands of Reassuring Parables about Things Humans Are Better Than Computers at Every Second.
-
2019-05-24-gpt2-poetry-yeatssecondcoming-500completions.txt
-
https://www.awanderingmind.blog/posts/2024-01-14-tao-te-ching-by-an-llm.html
-
https://x.com/HW
-
https://web.archive.org/web/20200209040154/https://decaut.org/situ/index.php/ttc-compilation/
-
2019-07-19-taotehching-ch1-1ksamples.txt
-
2019-07-21-gwern-gpt2-345m-taotehching-all.tar.xz
-
2019-07-21-taotehching-all-1ksamples.txt
-
2019-07-22-gpt2-345m-taotehching-all-ch181.tar.xz
-
Release Strategies and the Social Impacts of Language Models
-
Swarm Training: We Demonstrate a New Technique to Train ML Models Using Dozens of Independent TPUs.
-
Shawwn/gpt-2: Code for the Paper "Language Models Are Unsupervised Multitask Learners"
-
Pricing
-
Danbooru2019 Is a Large-Scale Anime Image Database With 3.69m+ Images Annotated With 108m+ Tags; It Can Be Useful for Machine Learning Purposes such as Image Recognition and Generation.
[Transclude the forward-link's context]
-
The Google SRE Handbook: Chapter 4—Service Level Objectives
-
HOGWILD!: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent
-
TensorFlow Research Cloud (TRC): Accelerate your cutting-edge machine learning research with free Cloud TPUs
-
GPT-1: Improving Language Understanding by Generative Pre-Training § Model specifications
-
ftfy: fixes text for you
-
Adafactor: Adaptive Learning Rates with Sublinear Memory Cost
-
U B U W E B :: Racter
-
Language Models Are Unsupervised Multitask Learners § Experiments
-
2019-12-13-gpt21.5b-poetry-samples-topp090.txt
-
2019-12-15-gpt21.5b-poetry-samples-topp090.txt
-
2019-12-16-gpt21.5b-poetry-samples-topp080.txt
-
2019-12-18-gpt21.5b-poetry-samples-topp080.txt
-
Greg Brockman: OpenAI and AGI
-
Figure F.1: Four Uncurated Completions from a Context Suggesting the Model Compose a Poem in the Style of Wallace Stevens With the Title ‘Shadows on the Way’
-
https://github.com/karpathy/char-rnn/issues/138
-
https://news.ycombinator.com/item?id=21335120
-
true_poetry: Poetry generator by GPT-2 with meter and rhyme constraints
-
MuZero: Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
-
Neural Text Generation with Unlikelihood Training
-
Do Massively Pretrained Language Models Make Better Storytellers?
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
-
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
-
Why Tool AIs Want to Be Agent AIs
-
Deep reinforcement learning from human preferences
-
CAN: Creative Adversarial Networks, Generating "Art" by Learning About Styles and Deviating from Style Norms
-
AlphaStar: Mastering the Real-Time Strategy Game StarCraft II
-
https://www.reddit.com/r/slatestarcodex/comments/b1b47h/gwerns_aigenerated_poetry/
-
https://news.ycombinator.com/item?id=19399467
-
https://news.ycombinator.com/item?id=21456403
-
Some Pretty Impressive Machine-Learning Generated Poetry Courtesy of GPT-2
-
Hark! from Those Shadowy Depths Thy Voice / Mournfully Echoes
-
On the Significance of Gwern’s Poem Generator
-
OpenAI’s New Language AI Is Available to Try Yourself
-
Generates Rhyming Poetry Using Huggingface GPT-2
-
A Hundred Visions and Revisions
-
RoBERTa: A Robustly Optimized BERT Pretraining Approach
-
Simonepri/lm-Scorer: 📃Language Model Based Sentences Scoring Library
-
How to Fine-Tune GPT-2 on Podcast Transcripts
-
These WWDC Boxed Lunches Aren't Real
-
https://web.archive.org/web/20220526054159/http://bkkaggle.github.io/blog/algpt2/2020/06/22/ALGPT2-part-1
-
https://web.archive.org/web/20210131134147/https://bkkaggle.github.io/blog/algpt2/2020/07/17/ALGPT2-part-2.html
-
The Average Fourth Grader Is a Better Poet Than…
-
The First Sally (A), Or, Trurl’s Electronic Bard
-
Seduced, Shaggy Samson Snored: The Fictional Machine That Generated Poems, and the Real People Who Had to Translate Them
-
Ramon Lull’s Thinking Machine
-
How to Build a State-Of-The-Art Conversational AI With Transfer Learning by Thomas Wolf
-
Computer Generated Foundation
-
https://www.reddit.com/r/SubSimulatorGPT2/comments/btfhks/what_is_rsubsimulatorgpt2/
-
A Chinese Room Writes a Sequel to Blindsight
-
How To Make Custom AI-Generated Text With GPT-2
-
Minimaxir/gpt-2-Keyword-Generation: Method to Encode Text for GPT-2 to Generate Text Based on Provided Keywords
-
Evaluation Metrics for Language Modeling
-
Lessons Learned from Building an AI Writing App
-
Excavate
-
Introducing Aspects of Creativity in Automatic Poetry Generation
-
Smart Vet: Autocompleting Sentences in Veterinary Medical Records
-
Deepfake Bot Submissions to Federal Public Comment Websites Cannot Be Distinguished from Human Submissions
-
This Word Does Not Exist [Github]
-
https://towardsdatascience.com/how-to-fine-tune-gpt-2-so-you-can-generate-long-form-creative-writing-7a5ae1314a61
-
This AI Poet Mastered Rhythm, Rhyme, and Natural Language to Write Like Shakespeare
-
Deep-speare: A Joint Neural Model of Poetic Language, Meter and Rhyme
-
Progressive Generation of Long Text
-
AdapterHub - 625 Adapters for 71 Text Tasks and 97 Languages
-
AdapterHub: A Framework for Adapting Transformers
-
Collaborative Storytelling with Large-scale Neural Language Models
-
Controllable Neural Text Generation
-
This Article Provides an Overview of Recent Methods to Fine-Tune Large Pre-Trained Language Models
-
Making Pre-trained Language Models Better Few-shot Learners
-
Prefix-Tuning: Optimizing Continuous Prompts for Generation
-
GPT Understands, Too
-
The Power of Scale for Parameter-Efficient Prompt Tuning
-
Entailment as Few-Shot Learner
-
Controllable Generation from Pre-trained Language Models via Inverse Prompting
-
https://gaotianyu.xyz/prompting/
-
Cutting Down on Prompts and Parameters: Simple Few-Shot Learning with Language Models
-
DART: Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners
-
PPT: Pre-trained Prompt Tuning for Few-shot Learning
-
Towards a Unified View of Parameter-Efficient Transfer Learning
-
2019-12-18-skylion-archiveofourown-fanfics-textscrape.tar.xz
-
https://archive.org/details/@entropy11235813
-
2020-01-14-gpt2-1558m-archiveofourownao3.tar.xz
-
AI Dungeon 2
-
2020-02-03-gpt21.5b-archiveofourownao3-model-510427-samples-topp090.txt
-
https://x.com/astraliteheart
-
Expanding the Frontiers of AI Creativity
-
😇A PyTorch Implementation of the DeepMoji Model: State-Of-The-Art Deep Learning Model for Analyzing Sentiment, Emotion, Sarcasm Etc
-
This Pony Does Not Exist
-
Yzhou359/MakeItTalk
-
2021-05-05-astraliteheart-purplesmartai-mylittleponygpt215b-twilightsparkledialogue.png
-
2021-05-05-astraliteheart-purplesmartai-mylittleponygpt215b-twilightsparkledialogue-torchmojiemotionalvoicecontrol.jpg
-
2021-05-05-astraliteheart-purplesmartai-mylittleponygpt215b-twilightsparkledialogue-voicedialogue.png
-
2021-05-05-astraliteheart-purplesmartai-mylittleponygpt215b-gpuloadgraph.png
-
End to End Agent Conversation Demo
-
My Little Pony: Friendship Is Magic Fanfiction
-
Library Genesis
-
Best Science Fiction (3506 Books)
-
The Best Fantasy Books
-
https://x.com/me_irl/status/1217818112957014022
-
2020-02-03-gpt21.5b-videogamewalkthrough-model-174925-samples-topp090.txt
-
OpenAI Text Generator GPT-2 Creates Video Game Walkthrough for ‘Most Tedious Game in History’
-
https://x.com/theshawwn/status/1212156603140648961
-
2025-03-08-2019-12-18-shawnpresser-gpt-2-117m-rdota2.tar.xz
-