Abstract
Class imbalance significantly affects the performance of machine learning and deep learning classifiers, especially in image recognition tasks where certain classes are underrepresented. Traditional synthetic oversampling methods, while helpful, often fail to address the complexities of real-world data with uneven class distributions. The first contribution of this study is the creation of artificially imbalanced datasets from CIFAR10 and SVHN datasets, designed to systematically evaluate classifier performance under varying degrees of class disparity. The second contribution is the introduction of the Clustering and Nearest Centroid Neighbour-based Synthetic Minority Oversampling (CLNCN-SMOTE) algorithm to resolve multi-class imbalance. The algorithm is an enhancement of traditional K-means SMOTE that incorporates a nearest centroid neighbour strategy. This method effectively generates more representative synthetic samples for the minority class, thereby reducing noise and mitigating overfitting issues. Building on this, the third contribution is the development of an enhanced framework that is different from traditional methods, as it integrates oversampling with self-supervised contrastive learning and attention mechanism to tackle multi-class imbalance. The proposed framework integrated the Convolutional Block Attention Module (CBAM) within the Momentum Contrast (MoCo) framework's encoder. The enhanced encoder was initially pre-trained using unlabelled images to refine its capability. Subsequently, the encoder was utilised to extract features from the training dataset and augment them using the CLNCN-SMOTE approach in the feature space. Finally, a Multi-layer Perceptron (MLP) classifier assessed the effectiveness of the entire framework. The proposed framework leverages contrastive learning to distinguish more effectively between features of different categories and employs an attention mechanism to optimise feature selection. This improves the classifier's ability to accurately recognise features of the minority class without degrading the majority class's performance. Experiments on several benchmark datasets, including CIFAR10, SVHN, Caltech-101, ImageNet-LT, and iNaturalist 2018, demonstrate significant improvements. The proposed framework achieves a macro-averaged F1-score (FM) of 73.17% and a macro-averaged geometric mean (GM) of 83.04% on CIFAR10, with FM reaching 79.61% and GM 88.28% on various SVHN datasets. Notably, it surpasses the MoCo-v2 framework by 5% in both FM and GM on the imbalanced SVHN and CIFAR10 datasets, achieving Top-1 accuracy of 68.67% on ImageNet-LT and 73.5% on iNaturalist 2018. In conclusion, the results underscore the framework's effectiveness in tackling multi-class image imbalance and highlight its significant practical applications in fields such as medical imaging, surveillance, and anomaly detection. Future studies could explore the integration of prototype-based contrastive learning methods for further enhancement.
Metadata
| Item Type: | Thesis (PhD) |
|---|---|
| Creators: | Creators Email / ID Num. Xiaoling, Gao UNSPECIFIED |
| Contributors: | Contribution Name Email / ID Num. Thesis advisor Mohd Rosli, Marshima UNSPECIFIED Thesis advisor Jamil, Nursuriati UNSPECIFIED |
| Subjects: | Q Science > Q Science (General) Q Science > Q Science (General) > Machine learning |
| Divisions: | Universiti Teknologi MARA, Shah Alam > Faculty of Computer and Mathematical Sciences |
| Programme: | Doctor of Philosophy (Computer Science) |
| Keywords: | Range-Controlled SMOTE (RCSMOTE), Synthetic Minority Over-sampling Technique (SMOTE), Self-supervised Learning (SSL) |
| Date: | October 2025 |
| URI: | https://ir.uitm.edu.my/id/eprint/132623 |
Download
132623.pdf
Download (16kB)
Digital Copy
Physical Copy
ID Number
132623
Indexing
