“GPT-2: 1.5B Release”, Irene Solaiman, Jack Clark, Miles Brundage2019-11-05 (; backlinks; similar)⁠:

As the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5b parameters) of GPT-2 along with code and model weights to facilitate detection of outputs of GPT-2 models. While there have been larger language models released since August, we’ve continued with our original staged release plan in order to provide the community with a test case of a full staged release process. We hope that this test case will be useful to developers of future powerful models, and we’re actively continuing the conversation with the AI community on responsible publication.

Our findings:

  1. Humans find GPT-2 outputs convincing.

  2. GPT-2 can be fine-tuned for misuse.

  3. Detection is challenging.

  4. We’ve seen no strong evidence of misuse so far.

  5. We need standards for studying bias.

Next steps: Our experience with GPT-2 over the past 9 months has given us valuable insight into the challenges and opportunities for creating responsible publication norms in AI. We’re continuing our work on this issue via participation in the Partnership on AI’s “Responsible Publication Norms for Machine Learning” project and discussions with our colleagues in the research community.