Claim Missing Document
Check
Articles

Found 6 Documents
Search
Journal : Journal of Electronics, Electromedical Engineering, and Medical Informatics

A Comparative Study of Machine Learning Methods for Baby Cry Detection Using MFCC Features Riadi, Putri Agustina; Faisal, Mohammad Reza; Kartini, Dwi; Nugroho, Radityo Adi; Nugrahadi, Dodon Turianto; Magfira, Dike Bayu
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 1 (2024): January
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i1.350

Abstract

The vocalization of infants, commonly known as baby crying, represents one of the primary means by which infants effectively communicate their needs and emotional states to adults. While the act of crying can yield crucial insights into the well-being and comfort of a baby, there exists a dearth of research specifically investigating the influence of the audio range within a baby cry on research outcomes. The core problem of research is the lack of research on the influence of audio range on baby cry classification on machine learning. The purpose of this study is to ascertain the impact of the duration of an infant’s cry on the outcomes of machine learning classification and to gain knowledge regarding the accuracy of results F1 score obtained through the utilization of the machine learning method. The contribution is to enrich an understanding of the application of classification and feature selection in audio datasets, particulary in the context of baby cry audio. The utilized dataset, known as donate-a-cry-corpus, encompasses five distinct data classes and possesses a duration of seven seconds. The employed methodology consists of the spectrogram technique, cross-validation for data partitioning, MFCC feature extraction with 10, 20, and 30 coefficients, as well as machine learning models including Support Vector Machine, Random Forest, and Naïve Bayes. The findings of this study reveal that the Random Forest model achieved an accuracy of 0.844 and an F1 score of 0.773 when 10 MFCC coefficients were utilized and the optimal audio range was set at six seconds. Furthermore, the Support Vector Machine model with an RBF kernel yielded an accuracy of 0.836 and an F1 score of 0.761, while the Naïve Bayes model achieved an accuracy 0.538 and F1 score of 0.539. Notably, no discernible differences were observed when evaluating the Support Vector Machine and Naïve Bayes methods across the 1-7 second time trial. The implication of this research is to establish a foundation for the advancement of premature illness identification techniques grounded in the vocalizations of infants, thereby facilitating swifter diagnostic processes for pediatric practitioners.
Comparative Study of Various Hyperparameter Tuning on Random Forest Classification With SMOTE and Feature Selection Using Genetic Algorithm in Software Defect Prediction Suryadi, Mulia Kevin; Herteno, Rudy; Saputro, Setyo Wahyu; Faisal, Mohammad Reza; Nugroho, Radityo Adi
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 2 (2024): April
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i2.375

Abstract

Software defect prediction is necessary for desktop and mobile applications. Random Forest defect prediction performance can be significantly increased with the parameter optimization process compared to the default parameter. However, the parameter tuning step is commonly neglected. Random Forest has numerous parameters that can be tuned, as a result manually adjusting parameters would diminish the efficiency of Random Forest, yield suboptimal results and it will take a lot of time. This research aims to improve the performance of Random Forest classification by using SMOTE to balance the data, Genetic Algorithm as selection feature, and using hyperparameter tuning to optimize the performance. Apart from that, it is also to find out which hyperparameter tuning method produces the best improvement on the Random Forest classification method. The dataset used in this study is NASA MDP which included 13 datasets. The method used contains SMOTE to handle imbalance data, Genetic Algorithm feature selection, Random Forest classification, and hyperparameter tuning methods including Grid Search, Random Search, Optuna, Bayesian (with Hyperopt), Hyperband, TPE and Nevergrad. The results of this research were carried out by evaluating performance using accuracy and AUC values. In terms of accuracy improvement, the three best methods are Nevergrad, TPE, and Hyperband. In terms of AUC improvement, the three best methods are Hyperband, Optuna, and Random Search. Nevergrad on average improves accuracy by about 3.9% and Hyperband on average improves AUC by about 3.51%. This study indicates that the use of hyperparameter tuning improves Random Forest performance and among all the hyperparameter tuning methods used, Hyperband has the best hyperparameter tuning performance with the highest average increase in both accuracy and AUC. The implication of this research is to increase the use of hyperparameter tuning in software defect prediction and improve software defect prediction performance.
Optimizing Software Defect Prediction Models: Integrating Hybrid Grey Wolf and Particle Swarm Optimization for Enhanced Feature Selection with Popular Gradient Boosting Algorithm Angga Maulana Akbar; Herteno, Rudy; Saputro, Setyo Wahyu; Faisal, Mohammad Reza; Nugroho, Radityo Adi
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 2 (2024): April
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i2.388

Abstract

Software defects, also referred to as software bugs, are anomalies or flaws in computer program that cause software to behave unexpectedly or produce incorrect results. These defects can manifest in various forms, including coding errors, design flaws, and logic mistakes, this defect have the potential to emerge at any stage of the software development lifecycle. Traditional prediction models usually have lower prediction performance. To address this issue, this paper proposes a novel prediction model using Hybrid Grey Wolf Optimizer and Particle Swarm Optimization (HGWOPSO). This research aims to determine whether the Hybrid Grey Wolf and Particle Swarm Optimization model could potentially improve the effectiveness of software defect prediction compared to base PSO and GWO algorithms without hybridization. Furthermore, this study aims to determine the effectiveness of different Gradient Boosting Algorithm classification algorithms when combined with HGWOPSO feature selection in predicting software defects. The study utilizes 13 NASA MDP dataset. These dataset are divided into testing and training data using 10-fold cross-validation. After data is divided, SMOTE technique is employed in training data. This technique generates synthetic samples to balance the dataset, ensuring better performance of the predictive model. Subsequently feature selection is conducted using HGWOPSO Algorithm. Each subset of the NASA MDP dataset will be processed by three boosting classification algorithms namely XGBoost, LightGBM, and CatBoost. Performance evaluation is based on the Area under the ROC Curve (AUC) value. Average AUC values yielded by HGWOPSO XGBoost, HGWOPSO LightGBM, and HGWOPSO CatBoost are 0.891, 0.881, and 0.894, respectively. Results of this study indicated that utilizing the HGWOPSO algorithm improved AUC performance compared to the base GWO and PSO algorithms. Specifically, HGWOPSO CatBoost achieved the highest AUC of 0.894. This represents a 6.5% increase in AUC with a significance value of 0.00552 compared to PSO CatBoost, and a 6.3% AUC increase with a significance value of 0.00148 compared to GWO CatBoost. This study demonstrated that HGWOPSO significantly improves the performance of software defect prediction. The implication of this research is to enhance software defect prediction models by incorporating hybrid optimization techniques and combining them with gradient boosting algorithms, which can potentially identify and address defects more accurately
A Comparative Analysis of Polynomial-fit-SMOTE Variations with Tree-Based Classifiers on Software Defect Prediction Nur Hidayatullah, Wildan; Herteno, Rudy; Reza Faisal, Mohammad; Adi Nugroho, Radityo; Wahyu Saputro, Setyo; Akhtar, Zarif Bin
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 3 (2024): July
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i3.455

Abstract

Software defects present a significant challenge to the reliability of software systems, often resulting in substantial economic losses. This study examines the efficacy of polynomial-fit SMOTE (pf-SMOTE) variants in combination with tree-based classifiers for software defect prediction, utilising the NASA Metrics Data Program (MDP) dataset. The research methodology involves partitioning the dataset into training and test subsets, applying pf-SMOTE oversampling, and evaluating classification performance using Decision Trees, Random Forests, and Extra Trees. Findings indicate that the combination of pf-SMOTE-star oversampling with Extra Tree classification achieves the highest average accuracy (90.91%) and AUC (95.67%) across 12 NASA MDP datasets. This demonstrates the potential of pf-SMOTE variants to enhance classification effectiveness. However, it is important to note that caution is warranted regarding potential biases introduced by synthetic data. These findings represent a significant advancement over previous research endeavors, underscoring the critical role of meticulous algorithm selection and dataset characteristics in optimizing classification outcomes. Noteworthy implications include advancements in software reliability and decision support for software project management. Future research may delve into synergies between pf-SMOTE variants and alternative classification methods, as well as explore the integration of hyperparameter tuning to further refine classification performance.
Optimization of Backward Elimination for Software Defect Prediction with Correlation Coefficient Filter Method Muhammad Noor; Radityo Adi Nugroho; Setyo Wahyu Saputro; Rudy Herteno; Friska Abadi
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 4 (2024): October
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i4.466

Abstract

Detecting software defects is a crucial step for software development not only to reduce cost and save time, but also to mitigate more costly losses. Backward Elimination is one method for detecting software defects. Notably Backward Elimination may remove features that may later become significant to the outcome affecting the performance of Backward Elimination. The aim of this study is to improve Backward Elimination performance. In this study, several features were selected based on their correlation coefficient, with the selected feature applied to improve Backward Elimination final model performance. The final model was validated using cross validation with Naïve Bayes as the classification method on the NASA MDP dataset to determine the accuracy and Area Under the Curve (AUC) of the final model. Using top 10 correlation feature and Backward Elimination achieve an average result of 86.6% accuracy and 0.797 AUC, while using top 20 correlation feature and Backward Elimination achieved an average result of 84% accuracy and 0.812 AUC. Compare to using Backward Elimination and Naïve Bayes respectively the improvement using top 10 correlation feature as follows: AUC:1.52%, 13.53% and Accuracy: 13%, 12.4% while the improvement using top 20 correlation feature as follows: AUC:3.43%, 15.66% and Accuracy: 10.4%, 9.8%. Results showed that selecting the top 10 and top 20 feature based on its correlation before using Backward Elimination have better result than only using Backward Elimination. This result shows that combining Backward Elimination with correlation coefficient feature selection does improve Backward Elimination’s final model and yielding good results for detecting software defects.
The Comparison of Extreme Machine Learning and Hidden Markov Model Algorithm in Predicting The Recurrence Of Differentiated Thyroid Cancer Using SMOTE Aida, Nor; Saragih, Triando Hamonangan; Kartini, Dwi; Nugroho, Radityo Adi; Nugrahadi, Dodon Turianto
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 4 (2024): October
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i4.467

Abstract

Differentiated thyroid cancer is the most common type of thyroid cancer; the types in this category are papillary, follicular, and hurthel cell carcinoma. Up to 20% of DTCs will experience recurrence, although this figure reduces to 5% in low-risk patients. There is still little research on thyroid cancer prediction using a machine learning approach, especially the prediction recurrence of DTCs. This research aims to compare the performance of the Extreme Learning Machine and the Hidden Markov Model using SMOTE in predicting the recurrence of DTCs. The dataset used in this research is differentiated thyroid cancer recurrence from Kaggle. This research methodology comprises preprocessing, data sharing, SMOTE, ELM and HMM modeling algorithms, and evaluation. ELM with SMOTE gets the best results at a ratio of 90:10 with 35 hidden neurons that get an accuracy value of 1.00, precision 1.00, recall 1.00, and AUC 1.00. ELM modeling gets the best results at a ratio of 90:10 with 45 hidden neurons that get an accuracy value of 1.00, precision 1.00, recall 1.00, and AUC 1.00. HMM modeling gets the best value at a ratio of 70:30 with two hidden states and two iterations, which get an accuracy value of 0.8696, precision 0.8696, recall 0.7944, and AUC 0.9575. Last, HMM modeling with SMOTE gets the best results at a ratio of 60:40 with two hidden states and two iterations, with an accuracy value of 0.8696, precision of 0.8832, recall of 0.7848, and AUC of 0.9174. Based on the results of this study, it can be concluded that ELM with SMOTE gets the best performance, followed by ELM without SMOTE, HMM without SMOTE, and finally, HMM with SMOTE. The implication is that ELM with SMOTE can produce high accuracy in predicting the recurrence of DTCs.
Co-Authors Abdul Gafur Adi Mu'Ammar, Rifqi Adin Nofiyanto, Adin Ahmad Bahroini Ahmad Juhdi Ahmad Rusadi Aida, Nor Akhtar, Zarif Bin Alamudin, Muhammad Faiq Andi Farmadi Andi Farmadi Andi Farmadi Angga Maulana Akbar Arie Sapta Nugraha Arie Sapta Nugraha Aryanti, Agustia Kuspita Athavale, Vijay Anant Aylwin Al Rasyid Bayu Hadi Sudrajat Dendy Fadhel Adhipratama Dendy Deni Kurnia Dike Bayu Magfira, Dike Bayu Dodon Turianto Nugrahadi Dwi Kartini Dwi Kartini, Dwi Efendi Mohtar Emma Andini Erdi, Muhammad Faisal, Mohammad Reza Fatma Indriani Fauzan Luthfi, Achmad Fenny Winda Rahayu Fhadilla Muhammad Friska Abadi Friska Abadi Hanif Rahardian Herteno, Rudy Irwan Budiman Irwan Budiman Itqan Mazdadi, Muhammad Ivan Sitohang Maya Yusida Muhammad Angga Wiratama Muhammad Azmi Adhani Muhammad Fikri Muhammad Itqan Mazdadi Muhammad Latief Saputra Muhammad Noor Muhammad Reza Faisal, Muhammad Reza Muhammad Rizky Adriansyah Muhammad Rusli Muhammad Syahriani Noor Basya Basya Muhammad Zaien Muliadi Muliadi Muliadi Aziz Muliadi Muliadi Muliadi Muliadi Nur Hidayatullah, Wildan Nur Ridha Apriyanti Oni Soesanto Pratama, Muhammad Yoga Adha Putri, Nitami Lestari Rahmat Ramadhani Raidra Zeniananto Reina Alya Rahma Reza Faisal, Mohammad Riadi, Putri Agustina Rinaldi Rizal, Muhammad Nur Rizky Ananda, Muhammad Rozaq, Hasri Akbar Awal Rudy Herteno Rudy Herteno Rudy Herteno Salsha Farahdiba Saputro, Setyo Wahyu Saragih, Triando Hamonangan Sarah Monika Nooralifa Septiadi Marwan Annahar Setyo Wahyu Saputro Siena, Laifansan Suci Permata Sari Suryadi, Mulia Kevin Sutan Takdir Alam Wahyu Caesarendra Wahyu Ramadansyah Wahyu Saputro, Setyo Zaini Abdan