Bibliography (9):

  1. https://x.com/NolanoOrg/status/1634027966651834370

  2. GPT-3: Language Models are Few-Shot Learners

  3. https://github.com/NolanoOrg/llama-int4-quant/

  4. https://github.com/qwopqwop200/GPTQ-for-LLaMa

  5. GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers

  6. FLAN: Scaling Instruction-Finetuned Language Models

  7. EleutherAI/gpt-Neo: An Implementation of Model Parallel GPT-2 and GPT-3-Style Models Using the Mesh-Tensorflow Library.