“ davidjl” is one of the many “glitch tokens” identified by Jessica Rumbelow and Matthew Watkins of SERI-MATS as producing hallucinations in GPT-2, -3, and -3.5.
Most of these no longer produce hallucinations in GPT-4, but “ davidjl” still does.
lesswrong.com/posts/aPeJE8bS…
Jun 8, 2023 · 12:10 AM UTC
More examples from @SoC_trilogy:
Someone else found this shortly after GPT-4 was released, but I've not seen any followup. At least one of the GPT-2/3 glitch tokens, " davidjl", also causes GPT-4 to glitch. Given an entirely new token set, how did this one slip through the net? Any guesses?
@repligate @goodside
To clarify: The issue isn't that it can't say the string, it's that it can't read it. If you break up the " davidjl" into two or more substrings it's able to say it by concatenating them. But if you literally write “ davidjl” (with a leading space) it can’t read it.
Important correction, with apologies to davidjl on YouTube — the source of our problems is Reddit, as usual:
If you liked this example and want to learn more about tokenization, see this great post by @simonw (who you should follow!):
Understanding GPT tokenizers: I wrote about how the tokenizers used by the various GPT models actually work, including an interactive tool for experimenting with their output simonwillison.net/2023/Jun/8…