Frans Richard Kodong
Jurusan Teknik Informatika UPN "Veteran" Yogyakarta

Published : 2 Documents Claim Missing Document
Claim Missing Document
Check
Articles

Found 2 Documents
Search

Comparison of K-Nearest Neighbor and Naïve Bayes algorithms for hoax classification in Indonesian health news Pratomo, Awang Hendrianto; Rachmad, Faiz; Kodong, Frans Richard
Bulletin of Social Informatics Theory and Application Vol. 8 No. 2 (2024)
Publisher : Association for Scientific Computing Electrical and Engineering

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.31763/businta.v8i2.796

Abstract

The categorization of health-related hoaxes is paramount in determining if they report facts. This paper analyzes the accuracy of the K-Nearest Neighbor (KNN) and the Naïve Bayes Classifier as two algorithms for health news hoaxes classification. Text mining was employed by feature extraction employing the TF-IDF method from the news headlines to classify the clusters. A prototype model was used to develop the system. Models assessment included confusion matrices and k-fold cross-validation. K=3 KNN model attained an average accuracy of 82.91%, precision of 85.3% and recall of 79.38% with no predictors included. The best performance was recorded for using the Naive Bayes model at fixation of K=3 KNN model at an average accuracy of 86.42%, precision level of 88.10% and recall high of 84.05%. These findings suggest that the KNN surfaces in the last model level rather than in the absence of the Naive Bayes model concerning classifying the hoax position of health news visible through the confusion evaluative matrix. Although related studies have been conducted in the past, this study is dissimilar in terms of its preprocessing methods, size of the data, and outcomes. The dataset consists of 1219 hoaxes labelled and 1227 facts labelled news headlines
Implementation of Mel-Frequency Cepstral Coefficient As Feature Extraction Method On Speech Audio Data Marbun, Andre Julio; Heriyanto; Kodong, Frans Richard
Telematika Vol 21 No 3 (2024): Edisi Oktober 2024
Publisher : Jurusan Informatika

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.31315/telematika.v21i3.12339

Abstract

Sounds cannot be directly processed by machines without a feature extraction process being carried out first. Currently, there are so many choices of feature extraction methods that can be used, so determining the right feature extraction method is not easy. One method of feature extraction on sound signals that is often used is Mel-Frequency Cepstral Coefficient (MFCC). MFCC has a working principle that resembles the human hearing system, which causes it to be widely used in various tasks related to recognition based on sound signals. This research will use the MFCC method to extract characteristics on voice signals and Support Vector Machine as a method of emotion classification on the RAVDESS dataset. MFCC consists of several stages, namely Pre-emphasize, Frame Blocking, Windowing, Fast Fourier Transform, Mel-Scaled Filterbank, Discrete Cosine Transform, and Cepstral Liftering. The type of test design that will be carried out in this research is parameter tuning. Parameter tuning is carried out with the aim of obtaining parameters that produce the best accuracy in the machine learning model. The parameters that will be tuned include the α value in the Pre-Emphasis process, frame length and overlap length in the Frame Blocking process, the number of mel filters in the Mel-Scaled Filterbank process, the number of cepstral coefficients in the Discrete Cosine Transform process and the C value in SVM. The best accuracy in males of 85.71% was obtained with a combination of filter parameter pre-emphasize of 0.95, frame length of 0.023 ms, overlap of adjacent frames of 40%, number of mel filters in the mel-scaled filterbank process of 24 mel, number of cepstral coefficient of 24 coefficient and the value of 'C' in SVM of 0.01. The best accuracy in women of 92.21% was obtained with a combination of filter parameters pre-emphasize of 0.95, frame length of 0.023 ms, overlap of adjacent frames of 40%, the number of mel filters in the melscaled filterbank process of 24 mel, and the number of cepstral coefficient of 13 coefficient and 'C' value in SVM of 0.01. From the two test results of tuning parameters between men and women, there are similar parameter values in all test parameters, except for the number of cepstral coefficients. The number of cepstral coefficient in men is 24 coefficient while the number of cepstral coefficient in women is 13 coefficient. Based on the research conducted, there are the following conclusions, the combination of MFCC and SVM methods can be used for emotion classification based on input data in the form of voice intonation with an accuracy of 85.71% in men and 92.21% in women. The difference in accuracy obtained between male and female models is due to the different data used. Male models are trained with male voice data and female models are trained with female voice data, this is done because men and women have different voice frequency ranges.