Machine Learning — Hard
Key points
- SimCLR uses contrastive learning to improve self-supervised representation learning
- Positive pairs are created by applying different augmentations to the same image
- NT-Xent loss is used to maximize agreement between representations
- Dissimilar samples are pushed apart in the embedding space
Ready to go further?
Related questions
