Angga Adiansya
Unknown Affiliation

Published : 2 Documents Claim Missing Document
Claim Missing Document
Check
Articles

Found 2 Documents
Search

The Implementation of a Logistic Regression Algorithm and Gradient Boosting Classifier for Predicting Telco Customer Churn Angga Adiansya; Zaenal Abidin
Pixel :Jurnal Ilmiah Komputer Grafis Vol 17 No 1 (2024): Vol 17 No 1 (2024): Jurnal Ilmiah Komputer Grafis
Publisher : UNIVERSITAS STEKOM

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.51903/pixel.v17i1.2006

Abstract

This research aims to predict customer churn in a telecommunications company using Logistic Regression (LR) and Gradient Boosting Classifier (GBC) algorithms. Customer churn poses a significant challenge as acquiring new customers is costlier than retaining existing ones. The dataset from Kaggle comprises 7043 records and 21 attributes. The process includes data pre-processing, cleaning, transformation, and normalization using a Min-Max Scaler. The data is split into features (X) and target (y), then divided into training and testing sets with an 80:20 ratio. Both models were trained and evaluated using a confusion matrix. Results show that the GBC model outperforms the LR model, with an accuracy of 83% compared to LR's 81%. This study demonstrates the effectiveness of GBC in predicting customer churn.
The Implementation of a Logistic Regression Algorithm and Gradient Boosting Classifier for Predicting Telco Customer Churn Angga Adiansya; Zaenal Abidin
Pixel :Jurnal Ilmiah Komputer Grafis Vol. 17 No. 1 (2024): Pixel :Jurnal Ilmiah Komputer Grafis dan Ilmu Komputer
Publisher : UNIVERSITAS STEKOM

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.51903/pixel.v17i1.2006

Abstract

This research aims to predict customer churn in a telecommunications company using Logistic Regression (LR) and Gradient Boosting Classifier (GBC) algorithms. Customer churn poses a significant challenge as acquiring new customers is costlier than retaining existing ones. The dataset from Kaggle comprises 7043 records and 21 attributes. The process includes data pre-processing, cleaning, transformation, and normalization using a Min-Max Scaler. The data is split into features (X) and target (y), then divided into training and testing sets with an 80:20 ratio. Both models were trained and evaluated using a confusion matrix. Results show that the GBC model outperforms the LR model, with an accuracy of 83% compared to LR's 81%. This study demonstrates the effectiveness of GBC in predicting customer churn.