Bibliography (6):

  1. Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context

  2. https://research.facebook.com/publications/llama-open-and-efficient-foundation-language-models/

  3. LLaMa-1: Open and Efficient Foundation Language Models

  4. https://openai.com/index/gpt-4-research/