Skip to content

This project employs advanced SMOTE variants to effectively work on imbalanced classification challenges in machine learning datasets.

Notifications You must be signed in to change notification settings

Leg0shii/SMOTE-Comparison

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

Handling Imbalanced Datasets with SMOTE Variants

Overview

This project explores the application of advanced SMOTE (Synthetic Minority Over-sampling Technique) algorithms—Borderline-SMOTE, Borderline-SMOTE2, and CURE SMOTE—to effectively address classification challenges in imbalanced datasets. By implementing machine learning models such as Random Forest and K-Nearest Neighbors, this study evaluates the performance improvements achieved through these oversampling techniques across three diverse datasets:

  • Mammography Dataset
  • Credit Card Fraud Dataset
  • ParkourMaker Dataset

Results

This project achieved the highest grade of 1.0 for its comprehensive approach to handling imbalanced datasets and the effective implementation of SMOTE variants. The application of CURE SMOTE, in particular, demonstrated significant improvements in minority class prediction accuracy, especially within the credit card fraud dataset when combined with the Random Forest algorithm.

About

This project employs advanced SMOTE variants to effectively work on imbalanced classification challenges in machine learning datasets.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published