Claim Missing Document
Check
Articles

Found 3 Documents
Search

Implementation of SMOTE and whale optimization algorithm on breast cancer classification using backpropagation Erlianita, Noor; Itqan Mazdadi, Muhammad; Saragih, Triando Hamonangan; Reza Faisal, Mohammad; Muliadi
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 5 No 4 (2023): October
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v5i4.334

Abstract

Breast cancer, which is characterized by uncontrolled cell growth, is the primary cause of mortality among women worldwide. The unchecked proliferation of cells leads to the formation of a mass or tumor. Generally, the absence of timely and efficient treatment contributes to this phenomenon. To prevent breast cancer, one of the strategies involves the classification of malignant and non-malignant types. For this particular investigation, the Breast Cancer Wisconsin dataset (original) comprising 699 instances with 11 classes and 1 target attribute was utilized. Synthetic Minority Oversampling (SMOTE) was employed to balance the dataset, with the Backpropagation classification algorithm and the Whale Optimization Algorithm (WOA) serving as optimization techniques. The main objectives of this study were to analyze the impact of the backpropagation method and SMOTE, examine the effect of the backpropagation method in conjunction with WOA, and assess the outcome of using the backpropagation method and SMOTE after incorporating WOA. The evaluation of the study's findings was performed using a confusion matrix and the Area Under the Curve (AUC) metric. The research outcomes based on the application of backpropagation yielded an accuracy rate of 96%, precision of 94%, recall of 95%, and an AUC of 96%. Subsequently, upon implementing SMOTE and WOA, the performance of the backpropagation method improved, resulting in an accuracy rate of 99%, precision of 97%, recall of 97%, and an AUC of 98%. This notable enhancement in performance suggests that the utilization of SMOTE and WOA effectively enhances accuracy. However, it is important to note that the observed improvements are relatively modest in nature.
Seleksi Fitur dengan Particle Swarm Optimization pada Klasifikasi Penyakit Parkinson Menggunakan XGBoost Kurnia, Deni; Itqan Mazdadi, Muhammad; Kartini, Dwi; Adi Nugroho, Radityo; Abadi, Friska
Jurnal Teknologi Informasi dan Ilmu Komputer Vol 10 No 5: Oktober 2023
Publisher : Fakultas Ilmu Komputer, Universitas Brawijaya

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.25126/jtiik.2023107252

Abstract

Penyakit Parkinson merupakan gangguan pada sistem saraf pusat yang mempengaruhi sistem motorik. Diagnosis penyakit ini cukup sulit dilakukan karena gejalanya yang serupa dengan penyakit lain. Saat ini diagnosa dapat dilakukan menggunakan machine learning dengan memanfaatkan rekaman suara pasien. Fitur yang dihasilkan dari ekstraksi rekaman suara tersebut relatif cukup banyak sehingga seleksi fitur perlu dilakukan untuk menghindari memburuknya kinerja sebuah model. Pada penelitian ini, Particle Swarm Optimization digunakan sebagai seleksi fitur, sedangkan XGBoost akan digunakan sebagai model klasifikasi. Selain itu model juga akan diterapkan SMOTE untuk mengatasi masalah ketidakseimbangan kelas data dan hyperparameter tuning pada XGBoost untuk mendapatkan hyperparameter yang optimal. Hasil pengujian menunjukkan bahwa nilai AUC pada model dengan seleksi fitur tanpa SMOTE dan hyperparameter tuning adalah 0,9325, sedangkan pada model tanpa seleksi fitur hanya mendapat nilai AUC sebesar 0,9250. Namun, ketika kedua teknik SMOTE dan hyperparameter tuning digunakan bersamaan, penggunaan seleksi fitur mampu memberikan peningkatan kinerja pada model. Model dengan seleksi fitur mendapat nilai AUC sebesar 0,9483, sedangkan model tanpa seleksi fitur hanya mendapat nilai AUC sebesar 0,9366.   Abstract   Parkinson's disease is a disorder of the central nervous system that affects the motor system. Diagnosis of this disease is quite difficult because the symptoms are similar to other diseases. Currently, diagnosis can be done using machine learning by utilizing patient voice recordings. The features generated from the extraction of voice recordings are relatively large, so feature selection needs to be done to avoid deteriorating the performance of a model. In this research, Particle Swarm Optimization is used as feature selection, while XGBoost will be used as a classification model. In addition, the model will also be applied SMOTE to overcome the problem of data class imbalance and hyperparameter tuning on XGBoost to get optimal hyperparameters. The test results show that the AUC value on the model with feature selection without SMOTE and hyperparameter tuning is 0.9325, while the model without feature selection only gets an AUC value of 0.9250. However, when both SMOTE and hyperparameter tuning techniques are used together, the use of feature selection is able to provide improved performance on the model. The model with feature selection gets an AUC value of 0.9483, while the model without feature selection only gets an AUC value of 0.9366.
Enhancing Software Defect Prediction: HHO-Based Wrapper Feature Selection with Ensemble Methods Fauzan Luthfi, Achmad; Herteno, Rudy; Abadi, Friska; Adi Nugroho, Radityo; Itqan Mazdadi, Muhammad; Athavale, Vijay Anant
Indonesian Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol. 7 No. 2 (2025): May
Publisher : Jurusan Teknik Elektromedik, Politeknik Kesehatan Kemenkes Surabaya, Indonesia

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/f2140043

Abstract

The growing complexity of data across domains highlights the need for effective classification models capable of addressing issues such as class imbalance and feature redundancy. The NASA MDP dataset poses such challenges due to its diverse characteristics and highly imbalanced classes, which can significantly affect model accuracy. This study proposes a robust classification framework integrating advanced preprocessing, optimization-based feature selection, and ensemble learning techniques to enhance predictive performance. The preprocessing phase involved z-score standardization and robust scaling to normalize data while reducing the impact of outliers. To address class imbalance, the ADASYN technique was employed. Feature selection was performed using Binary Harris Hawk Optimization (BHHO), with K-Nearest Neighbor (KNN) used as an evaluator to determine the most relevant features. Classification models including Random Forest (RF), Support Vector Machine (SVM), and Stacking were evaluated using performance metrics such as accuracy, AUC, precision, recall, and F1-measure. Experimental results indicated that the Stacking model achieved superior performance in several datasets, with the MC1 dataset yielding an accuracy of 0.998 and an AUC of 1.000. However, statistical significance testing revealed that not all observed improvements were meaningful; for example, Stacking significantly outperformed SVM but did not show a significant difference when compared to RF in terms of AUC. This underlines the importance of aligning model choice with dataset characteristics. In conclusion, the integration of advanced preprocessing and metaheuristic optimization contributes positively to software defect prediction. Future research should consider more diverse datasets, alternative optimization techniques, and explainable AI to further enhance model reliability and interpretability.