First of all, they specify a fast-learning "online" network, and a slow-learning, "target" one.
For a given sample the online network is trying to predict target network's embedding.
The catch?
They are using different augmentations!
This way they are taught to achieve consistent embeddings of observations across different ways of introducing noise.
https://arxiv.org/abs/2006.07733