CS 5330 · Assessment #2
Learning Without Labels
An interactive walkthrough of contrastive self-supervised learning via SimCLR. Explore how embedding spaces form, how augmentation creates positive pairs, and how the NT-Xent loss works.
Chen et al. 2020 — arXiv:2002.05709
Demo 02
Augmentation Playground
Toggle augmentations on/off and see how two different views of the same image form a positive pair. The model must learn that these two views — despite looking different — represent the same underlying image.
→
SimCLR key finding: Crop + Color Jitter together is the strongest combination. Each augmentation alone barely helps — their composition is what forces the encoder to learn meaningful structure.
Both views are a positive pair — same image, different augmentations. The encoder must produce similar embeddings for both.