Knowledge Engineering and Data Science
Vol 4, No 1 (2021)

Backpropagation Neural Network with Combination of Activation Functions for Inbound Traffic Prediction

Purnawansyah Purnawansyah (Department of informatics, Universitas Muslim Indonesia)
Haviluddin Haviluddin (Faculty of Engineering, Universitas Mulawarman, Kota Samarinda, Indonesia)
Herdianti Darwis (Universitas Muslim Indonesia)
Huzain Azis (Unknown)
Yulita Salim (Unknown)



Article Info

Publish Date
30 Jun 2021

Abstract

Predicting network traffic is crucial for preventing congestion and gaining superior quality of network services. This research aims to use backpropagation to predict the inbound level to understand and determine internet usage. The architecture consists of one input layer, two hidden layers, and one output layer. The study compares three activation functions: sigmoid, rectified linear unit (ReLU), and hyperbolic Tangent (tanh). Three learning rates: 0.1, 0.5, and 0.9 represent low, moderate, and high rates, respectively. Based on the result, in terms of a single form of activation function, although sigmoid provides the least RMSE and MSE values, the ReLu function is more superior in learning the high traffic pattern with a learning rate of 0.9. In addition, Re-LU is more powerful to be used in the first order in terms of combination. Hence, combining a high learning rate and pure ReLU, ReLu-sigmoid, or ReLu-Tanh is more suitable and recommended to predict upper traffic utilization

Copyrights © 2021






Journal Info

Abbrev

keds

Publisher

Subject

Computer Science & IT Engineering

Description

Knowledge Engineering and Data Science (2597-4637), KEDS, brings together researchers, industry practitioners, and potential users, to promote collaborations, exchange ideas and practices, discuss new opportunities, and investigate analytics frameworks on data-driven and knowledge base ...