
Self-supervised learning (SSL) is a powerful technique for learning from unlabeled data. By learning to remain invariant to applied data augmentations, methods such as SimCLR and MoCo can reach quality on par with supervised approaches. However, this invariance may be detrimental for solving downstream tasks that depend on traits affected by augmentations used during pretraining, such as color. In this paper, we propose to foster sensitivity to such characteristics in the representation space by modifying the projector network, a common component of self-supervised architectures. Specifically, we supplement the projector with information about augmentations applied to images. For the projector to take advantage of this auxiliary conditioning when solving the SSL task, the feature extractor learns to preserve the augmentation information in its representations. Our approach, coined Conditional Augmentation-aware Self-supervised Learning (CASSLE), is directly applicable to typical joint-embedding SSL methods regardless of their objective functions. Moreover, it does not require major changes in the network architecture or prior knowledge of downstream tasks. In addition to an analysis of sensitivity towards different data augmentations, we conduct a series of experiments, which show that CASSLE improves over various SSL methods, reaching state-of-the-art performance in multiple downstream tasks.
A short version of this paper appeared at the NeurIPS 2023 Workshop: Self-Supervised Learning - Theory and Practice (https://sslneurips23.github.io). The full paper was published (OA) in Knowledge-Based Systems (https://www.sciencedirect.com/science/article/pii/S0950705124012061)
FOS: Computer and information sciences, Computer Science - Machine Learning, contrastive learning, Computer Vision and Pattern Recognition (cs.CV), self-supervised learning, conditional models, Computer Science - Computer Vision and Pattern Recognition, augmentation-aware, projector, Machine Learning (cs.LG)
FOS: Computer and information sciences, Computer Science - Machine Learning, contrastive learning, Computer Vision and Pattern Recognition (cs.CV), self-supervised learning, conditional models, Computer Science - Computer Vision and Pattern Recognition, augmentation-aware, projector, Machine Learning (cs.LG)
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 4 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
