Stronger is not better: Better Augmentations in Contrastive Learning for Medical Image Segmentation
Self-supervised contrastive learning is among the recent representation learning methods that have shown performance gains in several downstream tasks including semantic segmentation. This paper evaluates strong data augmentation, one of the most important components for self-supervised contrastive learning’s improved performance. Strong data augmentation involves applying the composition of multiple augmentation techniques on images. Surprisingly, we find that the existing data augmentations do not always improve performance for semantic segmentation for medical images. We experiment with other augmentations that provide improved performance.
💡 Research Summary
This paper, titled “Stronger is not better: Better Augmentations in Contrastive Learning for Medical Image Segmentation,” presents a critical empirical investigation into the effectiveness of data augmentation strategies within self-supervised contrastive learning frameworks, specifically for the downstream task of medical image segmentation.
The core premise challenges a widely held assumption in self-supervised learning (SSL): that employing “strong” data augmentations—complex compositions of transformations like aggressive color distortion, Gaussian blur, and random cropping—is inherently beneficial for learning robust visual representations. While proven highly successful for general image classification on datasets like ImageNet, the authors hypothesize that such augmentations may not be optimal for domain-specific tasks like medical image segmentation.
The experimental setup utilizes the SimCLR framework for SSL pre-training and a U-Net for the downstream semantic segmentation task. The KVASIR-SEG dataset, containing 1000 endoscopic images of polyps with corresponding segmentation masks, serves as the testbed. The key variable is the augmentation pipeline used during the SimCLR pre-training phase. The authors compare the standard “strong” augmentations from the original SimCLR paper (random crop and resize, random color distortions, random Gaussian blur) against a set of “basic” augmentations (resize, rotation, horizontal flip).
Results are evaluated across multiple metrics (Dice Loss, IoU, F-Score, Recall, Precision) under varying conditions of batch size (8, 16, 32, 64) and weight initialization (random vs. ImageNet pre-trained). The findings are consistent and striking: models pre-trained using the simple basic augmentations consistently outperform or match those pre-trained with SimCLR’s strong augmentations across nearly all experimental configurations. For instance, with random initialization and a batch size of 32, the basic augmentation model achieved an IoU of 0.7756 compared to 0.7562 for the strong augmentation model. This performance advantage held even under conditions favorable to SimCLR, such as a larger batch size of 64 and ImageNet-initialized weights.
The authors provide a reasoned discussion for these counter-intuitive results. They argue that medical images possess distinct characteristics—consistent anatomical structures, limited color spectra, and subtle pathological textures—that are crucial for accurate segmentation. Overly aggressive augmentations, particularly color distortion and heavy blurring, may destroy or obfuscate these semantically important features, thereby hindering the model’s ability to learn meaningful, task-relevant representations. In contrast, simpler geometric augmentations like rotation and flipping preserve these critical features while still providing sufficient invariance for effective contrastive learning.
The paper concludes that the utility of an augmentation strategy is not defined by its “strength” but by its “suitability” to the target domain and task. Blindly applying augmentation policies developed for natural images to the medical imaging domain can be suboptimal. The work emphasizes the need for domain-aware design in SSL and calls for future research to theoretically understand why certain SSL components fail in specific contexts and to evaluate other elements of contrastive learning frameworks on medical data. This study serves as an important cautionary note and a step towards more nuanced, application-driven self-supervised learning methodologies.
Comments & Academic Discussion
Loading comments...
Leave a Comment