Separated and Independent Contrastive Semi-Supervised Learning for Imbalanced Datasets
Conventional semi-supervised learning (SSL) encounters challenges in effectively addressing issues associated with long-tail datasets, primarily stemming from imbalances within a dataset. Existing contrastive SSL methods typically rely on pseudo-labels generated from unlabeled data, which can be ina...
Saved in:
| Main Authors: | , , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/11039800/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Conventional semi-supervised learning (SSL) encounters challenges in effectively addressing issues associated with long-tail datasets, primarily stemming from imbalances within a dataset. Existing contrastive SSL methods typically rely on pseudo-labels generated from unlabeled data, which can be inaccurate and introduce confirmation bias toward majority classes. To address this issue, we propose Separated and Independent Contrastive Semi-Supervised Learning (SICSSL), which applies supervised contrastive learning separately and independently to labeled and unlabeled samples. We validate SICSSL on four public benchmark datasets: CIFAR-10-LT, CIFAR-100-LT, STL-10-LT, and ImageNet-127. The results demonstrate consistent performance improvements over existing contrastive learning approaches. Furthermore, SICSSL is modular and easily integrates with recent SSL frameworks, including Auxiliary Balanced Classifier (ABC) and Adaptive Consistency Regularizer (ACR), enabling improved robustness in imbalanced semi-supervised settings. Source code is available at <uri>https://github.com/dongyyyyy/SICSSL</uri> |
|---|---|
| ISSN: | 2169-3536 |