Claim Missing Document
Check
Articles

Found 6 Documents
Search

Journal Classification Using Cosine Similarity Method on Title and Abstract with Frequency-Based Stopword Removal  Piska Dwi Nurfadila; Aji Prasetya Wibawa; Ilham Ari Elbaith Zaeni; Andrew Nafalski
International Journal of Artificial Intelligence Research Vol 3, No 2 (2019): December 2019
Publisher : STMIK Dharma Wacana

Show Abstract | Download Original | Original Source | Check in Google Scholar | Full PDF (231.173 KB) | DOI: 10.29099/ijair.v3i2.99

Abstract

Classification of economic journal articles has been done using the VSM (Vector Space Model) approach and the Cosine Similarity method. The results of previous studies are considered to be less optimal because Stopword Removal was carried out by using a dictionary of basic words (tuning). Therefore, the omitted words limited to only basic words. This study shows the improved performance accuracy of the Cosine Similarity method using frequency-based Stopword Removal. The reason is because the term with a certain frequency is assumed to be an insignificant word and will give less relevant results. Performance testing of the Cosine Similarity method that had been added to frequency-based Stopword Removal was done by using K-fold Cross Validation. The method performance produced accuracy value for 64.28%, precision for 64.76 %, and recall for 65.26%. The execution time after pre-processing was 0, 05033 second.
Journal Unique Visitors Forecasting Based on Multivariate Attributes Using CNN Aderyan Reynaldi Fahrezza Dewandra; Aji Prasetya Wibawa; Utomo Pujianto; Agung Bella Putra Utama; Andrew Nafalski
International Journal of Artificial Intelligence Research Vol 6, No 2 (2022): Desember 2022
Publisher : International Journal of Artificial Intelligence Research

Show Abstract | Download Original | Original Source | Check in Google Scholar | Full PDF (379.839 KB) | DOI: 10.29099/ijair.v6i1.274

Abstract

Forecasting is needed in various problems, one of which is forecasting electronic journals unique visitors. Although forecasting cannot produce very accurate predictions, using the proper method can reduce forecasting errors. In this research, forecasting is done using the Deep Learning method, which is often used to process two-dimensional data, namely convolutional neural network (CNN). One-dimensional CNN comes with 1D feature extraction suitable for forecasting 1D time-series problems. This study aims to determine the best architecture and increase the number of hidden layers and neurons on CNN forecasting results. In various architectural scenarios, CNN performance was measured using the root mean squared error (RMSE). Based on the study results, the best results were obtained with an RMSE value of 2.314 using an architecture of 2 hidden layers and 64 neurons in Model 1. Meanwhile, the significant effect of increasing the number of hidden layers on the RMSE value was only found in Model 1 using 64 or 256 neurons.
Classification of Engineering Journals Quartile using Various Supervised Learning Models Nastiti Susetyo Fanany Putri; Aji Prasetya Wibawa; Harits Ar Rasyid; Anik Nur Handayani; Andrew Nafalski; Edinar Valiant Hawali; Jehad A.H. Hammad
ILKOM Jurnal Ilmiah Vol 15, No 1 (2023)
Publisher : Prodi Teknik Informatika FIK Universitas Muslim Indonesia

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.33096/ilkom.v15i1.1483.101-106

Abstract

In scientific research, journals are among the primary sources of information. There are quartiles or categories of quality in journals which are Q1, Q2, Q3, and Q4. These quartiles represent the assessment of journal. A classification machine learning algorithm is developed as a means in the categorization of journals. The process of classifying data to estimate an item class with an unknown label is called classification. Various classification algorithms, such as K-Nearest Neighbor (KNN), Naïve Bayes, and Support Vector Machine (SVM) are employed in this study, with several situations for exchanging training and testing data. Cross-validation with Confusion Matrix values of accuracy, precision, recall, and error classification is used to analyzed classification performance. The classifier with the finest accuracy rate is KNN with average accuracy of 70%, Naïve Bayes at 60% and SVM at 40%. This research suggests assumption that algorithms used in this article can approach SJR classification system.
Boosting and bagging classification for computer science journal Nastiti Susetyo Fanany Putri; Aji Prasetya Wibawa; Harits Ar Rasyid; Andrew Nafalski; Ummi Rabaah Hasyim
International Journal of Advances in Intelligent Informatics Vol 9, No 1 (2023): March 2023
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v9i1.985

Abstract

In recent years, data processing has become an issue across all disciplines. Good data processing can provide decision-making recommendations. Data processing is covered in academic data processing publications, including those in computer science. This topic has grown over the past three years, demonstrating that data processing is expanding and diversifying, and there is a great deal of interest in this area of study. Within the journal, groupings (quartiles) indicate the journal's influence on other similar studies. SCImago provides this category. There are four quartiles, with the highest quartile being 1 and the lowest being 4. There are, however, numerous differences in class quartiles, with different quartile values for the same journal in different disciplines. Therefore, a method of categorization is provided to solve this issue. Classification is a machine-learning technique that groups data based on the supplied label class. Ensemble Boosting and Bagging with Decision Tree (DT) and Gaussian Nave Bayes (GNB) were utilized in this study. Several modifications were made to the ensemble algorithm's depth and estimator settings to examine the influence of adding values on the resultant precision. In the DT algorithm, both variables are altered, whereas, in the GNB algorithm, just the estimator's value is modified. Based on the average value of the accuracy results, it is known that the best algorithm for computer science datasets is GNB Bagging, with values of 68.96%, 70.99%, and 69.05%. Second-place XGBDT has 67.75% accuracy, 67.69% precision, and 67.83 recall. The DT Bagging method placed third with 67.31 percent recall, 68.13 percent precision, and 67.30 percent accuracy. The fourth sequence is the XGBoost GNB approach, which has an accuracy of 67.07%, a precision of 68.85%, and a recall of 67.18%. The Adaboost DT technique ranks in the fifth position with an accuracy of 63.65%, a precision of 64.21 %, and a recall of 63.63 %. Adaboost GNB is the least efficient algorithm for this dataset since it only achieves 43.19 % accuracy, 48.14 % precision, and 43.2% recall. The results are still quite far from the ideal. Hence the proposed method for journal quartile inequality issues is not advised.
Evolving Conversations: A Review of Chatbots and Implications in Natural Language Processing for Cultural Heritage Ecosystems Tri Lathif Mardi Suryanto; Aji Prasetya Wibawa; Hariyono Hariyono; Andrew Nafalski
International Journal of Robotics and Control Systems Vol 3, No 4 (2023)
Publisher : Association for Scientific Computing Electronics and Engineering (ASCEE)

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.31763/ijrcs.v3i4.1195

Abstract

Chatbot technology, a rapidly growing field, uses Natural Language Processing (NLP) methodologies to create conversational AI bots. Contextual understanding is essential for chatbots to provide meaningful interactions. Still, to date chatbots often struggle to accurately interpret user input due to the complexity of natural language and diverse fields, hence the need for a Systematic Literature Review (SLR) to investigate the motivation behind the creation of chatbots, their development procedures and methods, notable achievements, challenges and emerging trends. Through the application of the PRISMA method, this paper contributes to revealing the rapid and dynamic progress in chatbot technology with NLP learning models, enabling sophisticated and human-like interactions on the trends observed in chatbots over the past decade. The results, from various fields such as healthcare, organization and business, virtual personalities, to education, do not rule out the possibility of being developed in other fields such as chatbots for cultural preservation while suggesting the need for supervision in the aspects of language comprehension bias and ethics of chatbot users. In the end, the insights gained from SLR have the potential to contribute significantly to the advancement of chatbots on NLP as a comprehensive field.
Comparative Performance of Transformer Models for Cultural Heritage in NLP Tasks Tri Lathif Mardi Suryanto; Aji Prasetya Wibawa; Hariyono Hariyono; Andrew Nafalski
Advance Sustainable Science Engineering and Technology Vol. 7 No. 1 (2025): November-January
Publisher : Science and Technology Research Centre Universitas PGRI Semarang

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26877/asset.v7i1.1211

Abstract

AI and Machine Learning are crucial in advancing technology, especially for processing large, complex datasets. The transformer model, a primary approach in natural language processing (NLP), enables applications like translation, text summarization, and question-answer (QA) systems. This study compares two popular transformer models, FlanT5 and mT5, which are widely used yet often struggle to capture the specific context of the reference text. Using a unique Goddess Durga QA dataset with specialized cultural knowledge about Indonesia, this research tests how effectively each model can handle culturally specific QA tasks. The study involved data preparation, initial model training, ROUGE metric evaluation (ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-Lsum), and result analysis. Findings show that FlanT5 outperforms mT5 on multiple metrics, making it better at preserving cultural context. These results are impactful for NLP applications that rely on cultural insight, such as cultural preservation QA systems and context-based educational platforms.