Bibliography (5):

  1. OPT: Open Pre-trained Transformer Language Models

  2. https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/

  3. https://github.com/microsoft/TransformerCompression