Bibliography (5):

  1. Attention Is All You Need

  2. Universal Transformers

  3. The LAMBADA dataset: Word prediction requiring a broad discourse context

  4. Wikipedia Bibliography:

    1. Turing completeness

    2. BLEU