This Author published in this journals
All Journal bit-Tech
Ciptaagung Firjat Ardine
Universitas Pembangunan Nasional Veteran Jawa Timur

Published : 1 Documents Claim Missing Document
Claim Missing Document
Check
Articles

Found 1 Documents
Search

Mobile Legends Match Outcome Prediction Based on Players Statistics Using CatBoost and XGBoost Ciptaagung Firjat Ardine; Eka Prakarsa Mandyartha; Achmad Junaidi
bit-Tech Vol. 8 No. 2 (2025): bit-Tech
Publisher : Komunitas Dosen Indonesia

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.32877/bt.v8i2.3259

Abstract

Mobile Legends: Bang Bang (MLBB) is a mobile-based Multiplayer Online Battle Arena (MOBA) game with a vast global community and professional ecosystem. Despite the extensive use of machine learning in desktop-based MOBAs such as Dota 2 and League of Legends, predictive modeling for MLBB remains underexplored. This study addresses this research gap by developing and comparing two advanced gradient boosting algorithms CatBoost and XGBoost for predicting match outcomes based on individual player statistics. The dataset, collected through web scraping from the official MPL Malaysia Season 14 website, comprises 1,430 player-level records representing professional-level competitive matches. Both models were trained and evaluated using 5-Fold Cross Validation to ensure stability and robustness. The results indicate that CatBoost achieved the highest predictive accuracy, with an average of 96.15%, outperforming XGBoost, which attained 94.75%. However, XGBoost exhibited exceptional computational efficiency, completing the prediction process 99.62% faster 0.76 seconds compared to CatBoost’s 3 minutes and 21 seconds. These findings highlight the trade-off between accuracy and processing speed in esports predictive modeling. The study demonstrates the potential of gradient boosting approaches for MLBB-specific analytics, providing a novel contribution to the limited body of research on mobile esports prediction. Accordingly, CatBoost is more suitable for analytical or strategic contexts where precision is essential, while XGBoost is better aligned with real-time predictive systems that demand rapid computation and scalability.