Claim Missing Document
Check
Articles

Found 34 Documents
Search

Comparison of K-Means and DBSCAN Algorithms for Customer Segmentation in E-commerce Paramita, Adi Suryaputra; Hariguna, Taqwa
Journal of Digital Market and Digital Currency Vol. 1 No. 1 (2024): Regular Issue June 2024
Publisher : Bright Publisher

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47738/jdmdc.v1i1.3

Abstract

Customer segmentation is crucial for e-commerce businesses to effectively target and engage specific customer groups. This study compares the effectiveness of two popular clustering algorithms, K-Means and DBSCAN, in segmenting e-commerce customers. The primary objective is to evaluate and contrast these algorithms to determine which provides more meaningful and actionable customer segments. The methodology involves analyzing a comprehensive e-commerce customer dataset, which includes various features such as customer ID, gender, age, city, membership type, total spend, items purchased, average rating, discount applied, days since last purchase, and satisfaction level. Initial data preprocessing steps include handling missing values, encoding categorical variables, and normalizing numerical features. Both K-Means and DBSCAN algorithms are implemented, and their performance is evaluated using metrics such as silhouette score, Davies-Bouldin index, and Calinski-Harabasz score. The results indicate that K-Means achieved a silhouette score of 0.546, a Davies-Bouldin index of 0.655, and a Calinski-Harabasz score of 552.9. In contrast, DBSCAN achieved a higher silhouette score of 0.680, a Davies-Bouldin index of 1.344, and a Calinski-Harabasz score of 1123.9. These findings suggest that while DBSCAN performs better in terms of silhouette score, indicating more distinctly separated clusters, its higher Davies-Bouldin index reflects fewer compact clusters. The discussion highlights that K-Means is suitable for applications requiring clear and well-defined segments of customers, as it produces balanced cluster sizes. DBSCAN, with its strength in identifying clusters of varying densities and handling noise, is more effective in detecting niche markets and unique customer behaviors. This study's findings have significant practical implications for e-commerce businesses looking to enhance their customer segmentation strategies. In conclusion, both K-Means and DBSCAN demonstrate their respective strengths and weaknesses in clustering the e-commerce customer dataset. The choice of algorithm should be based on the specific requirements of the segmentation task. Future research could explore hybrid methods combining the strengths of both algorithms and incorporate additional data sources for a more comprehensive analysis.
Predicting Player Performance in Valorant E-Sports using Random Forest Algorithm: A Data Mining Approach for Analyzing Match and Agent Data in Virtual Environments Paramita, Adi Suryaputra; Jusak, Jusak
International Journal Research on Metaverse Vol. 2 No. 4 (2025): Regular Issue December 2025
Publisher : Bright Publisher

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47738/ijrm.v2i4.39

Abstract

This study presents a data-driven approach to predict player performance in Valorant, an increasingly popular e-sport, using a Random Forest machine learning model. As e-sports continue to evolve within the metaverse, the need for strategic optimization and player selection has become critical. By analyzing a dataset containing player statistics from the Valorant Champion Tour (VCT), we aimed to predict player Rating, a key performance indicator. The dataset includes various metrics such as Kills Per Round, Average Combat Score (ACS), Clutch Success Ratio, and Kills:Deaths. After preprocessing the data, which involved handling missing values and feature engineering, the dataset was split into training and testing sets (80% and 20%, respectively). The Random Forest model, with 100 estimators and a maximum depth of 10, was trained on the processed data. The model's performance was evaluated using regression metrics such as Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and R-squared (R²). The results demonstrated that the model could predict player performance with a high degree of accuracy, with an R² value of 0.8831, meaning it explained 88.31% of the variance in player ratings. Furthermore, Kills Per Round emerged as the most significant feature, followed by Kill, Assist, Trade, Survive Ratio and Average Damage Per Round. These insights suggest that key metrics like kills and damage output are crucial for predicting player success. This study not only provides a comprehensive framework for predicting Valorant player performance but also demonstrates the potential of data mining in optimizing e-sports strategies. The findings contribute to the growing body of research on virtual gaming environments and offer actionable insights for teams in the metaverse, enabling data-driven decision-making to enhance performance and strategic outcomes.
Fault-Tolerant Telegram Bot Architecture for Odoo 14: Validated Production Reporting in Flexible Packaging Tarigan, Masmur; Paramita, Adi Suryaputra; Dewi, Deshinta Arrova
Jurnal Teknik Informatika (Jutif) Vol. 6 No. 6 (2025): JUTIF Volume 6, Number 6, Desember 2025
Publisher : Informatika, Universitas Jenderal Soedirman

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.52436/1.jutif.2025.6.6.5515

Abstract

In flexible-packaging manufacturing, manual reporting dramatically delays synchronization with the ERP — and that means operational  latency and traceability issues. The proposed work is the design, implementation, and validation of a fault-tolerant Telegram bot interconnected with Odoo 14 for six production departments. Our bot architecture that combines conversational workflows with schema-based validation and XML-RPC for slow, large payloads, enables accurate and  timely reporting. In a four-week pilot with 1,066 production entries, we achieved 98.7% field completeness and lowered reporting latency to less than 2 minutes. Manual  baselines received 75% more requests for corrections. At disconnected state, the layered middleware of the system abstracted retry logic and media ingestion. Both SDG 9 (Resilient infrastructure, including ) and SDG 12 (Continue to reduce production waste at source, including consumables) are connected to the work presented here which evidence the feasibility of automatic conversational interfaces with a computer in the manufacturing informatics domain, and provide pathways towards scalable digital transformation and sustainability in the small-to-medium industry sector.
Machine Learning-Based Fraud Detection in E-Commerce Transactions Evelyn, Evelyn; Paramita, Adi Suryaputra
International Journal of Informatics and Information Systems Vol 9, No 1: Regular Issue: January 2026
Publisher : International Journal of Informatics and Information Systems

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47738/ijiis.v9i1.295

Abstract

The rapid growth of e-commerce has heightened fraud risks, demanding advanced fraud detection solutions. This study evaluates five machine learning models Logistic Regression, SVM, KNN, Random Forest, and Gradient Boosting for detecting fraudulent transactions in e-commerce environments. The models were assessed based on accuracy, precision, recall, F1-score, ROC-AUC, and error-related indicators. Results indicate that ensemble-based models, particularly Gradient Boosting and Random Forest, consistently outperform linear models like Logistic Regression, achieving superior balance between precision and recall. Gradient Boosting emerged as the top performer, with the highest accuracy (0.9763), F1-score (0.9765), and ROC-AUC (0.9880), while maintaining a low false negative rate (4.38%). These findings suggest that machine learning models, particularly ensemble methods, provide robust and efficient fraud detection frameworks. The study emphasizes the importance of using recall and F1-score as primary metrics to balance fraud detection sensitivity and operational efficiency.