Do you remember how autoencoders can be used for pretraining on unlabeled data?

Now, imagine a method that can do it much better, much faster, without the pesky decoding part.

Welcome to a short thread on SimCLR 👇 (1/4)

SimCLR is an algorithm which depends on a common operation: augmenting the data.

However, in contrast to traditional methods, we don't use augmentation for pretending that we have more data. To the contrary! (2/4)

In SimCLR we are going to show our neural network two different augmentations of our sample and train it to recognize that this is one and the same thing.

This way we're training the network to look past the augmentations and focus on what's important on the image. (3/4)

Follow

If you look at the first picture, you may ask yourself a question - why do we need to treat the projection from h to z separately?

It turns out that it's a side effect of contrastive loss! z vectors contain only the bare minimum info needed to recognize the similarity. (4/4)

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.