Bibliography (4):

  1. Contrastive Representation Learning: A Framework and Review

  2. MAE: Masked Autoencoders Are Scalable Vision Learners

  3. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

  4. Wikipedia Bibliography:

    1. Data augmentation