Bibliography (9):

  1. Mind the Gap: Assessing Temporal Generalization in Neural Language Models § Scaling

  2. Recurrent Neural Network Based Language Model § Dynamic Evaluation

  3. Recurrent Neural Network Based Language Model

  4. Generating Sequences With Recurrent Neural Networks

  5. Dynamic Evaluation of Transformer Language Models

  6. https://arxiv.org/pdf/2102.01951.pdf#page=18&org=deepmind

  7. https://arxiv.org/pdf/2102.01951#page=5&org=deepmind

  8. Mind the Gap: Assessing Temporal Generalization in Neural Language Models § Dynamic Evaluation

  9. BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models