Abdullah, Zubaile
Unknown Affiliation

Published : 7 Documents Claim Missing Document
Claim Missing Document
Check
Articles

Found 7 Documents
Search

Object retrieval analysis on plastic bottle waste recycling-based image control using convex hull algorithm and autoregressive integrated moving average prediction method Marisa, Marisa; Azhar Ramli, Azizul; Fudzee, Mohd Farhan Md; Abdullah, Zubaile
International Journal of Electrical and Computer Engineering (IJECE) Vol 15, No 2: April 2025
Publisher : Institute of Advanced Engineering and Science

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.11591/ijece.v15i2.pp2055-2069

Abstract

In Indonesia, plastic garbage bottles are the most common sort of waste. Given that waste is expected to grow annually, managing plastic waste is a major challenge. The results of the study were achieved by comparing the reference, which was a collection of manually created contour images, with 50 sets of vortex images with different forms and vortex areas as experimental objects. The results indicate that the suggested approach reports a mean error of 2.84%, a correlation coefficient of 0.9965, and a root mean square error of 0.2903 when compared to the manual extraction method. These findings imply that the extract area determined by the procedure outlined in this research is more accurate and nearer to the actual values. The proposed method can therefore be used in place of the traditional process for investigating cooling parameters through manual testing. With measurement values mean absolute percentage error (MAPE)=121,842, mean absolute deviation (MAD)=20,140, and mean squared deviation (MSD)=776,712, the trend analysis of plastic bottles for autoregressive integrated moving average (ARIMA) modeling leads to the conclusion that the waste from plastic bottles will continue to rise annually and that efforts must be made to address this trend with knowledge and waste recycling technology. Plastic that is advantageous to industry and society.
Hybrid Logistic Regression Random Forest on Predicting Student Performance Rohman, Muhammad Ghofar; Abdullah, Zubaile; Kasim, Shahreen; Rasyidah, -
JOIV : International Journal on Informatics Visualization Vol 9, No 2 (2025)
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.62527/joiv.9.2.3972

Abstract

The research aims to investigate the effects of unbalanced data on machine learning, overcome imbalanced data using SMOTE oversampling, and improve machine learning performance using hyperparameter tuning. This study proposed a model that combines logistic regression and random forests as a hybrid logistic regression, random forest, and random search SV that uses SMOTE oversampling and hyperparameter tuning. The result of this study showed that the prediction model using the hybrid logistic regression, random forest, and random search SV that we proposed produces more effective performance than using logistic regression and random forest, with accuracy, precision, recall, and F1-score of 0.9574, 0.9665, 0.9576. This can contribute to a practical model to address imbalanced data classification based on data-level solutions for student performance prediction.
Comparative Analysis of Machine Learning Algorithms for Cross-Site Scripting (XSS) Attack Detection Hamzah, Khairatun Hisan; Osman, Mohd Zamri; Anthony, Tumusiime; Ismail, Mohd Arfian; Abdullah, Zubaile; Alanda, Alde
JOIV : International Journal on Informatics Visualization Vol 8, No 3-2 (2024): IT for Global Goals: Building a Sustainable Tomorrow
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.62527/joiv.8.3-2.3451

Abstract

Cross-Site Scripting (XSS) attacks pose a significant cybersecurity threat by exploiting vulnerabilities in web applications to inject malicious scripts, enabling unauthorized access and execution of malicious code. Traditional XSS detection systems often struggle to identify increasingly complex XSS payloads. To address this issue, this research evaluated the efficacy of Machine Learning algorithms in detecting XSS threats within online web applications. The study conducts a comprehensive comparative analysis of XSS attack detection using four prominent Machine Learning algorithms, which consist of Extreme Gradient Boosting (XGBoost), Random Forest (RF), K-Nearest Neighbors (KNN), and Support Vector Machine (SVM). This research utilizes a comparative methodology to assess the selected Machine Learning algorithms by analyzing their performance metrics, including confusion matrix, 10-fold cross-validation, and assessment of training time to thoroughly evaluate the models. By exploring dataset characteristics and evaluating the performance metrics of each selected algorithm, the study determined the most robust Machine Learning solution for XSS detection. Results indicate that Random Forest is the top performer, achieving 99.93% accuracy and balanced metrics across all criteria evaluated. These findings will significantly enhance web application security by providing reliable defenses against evolving XSS threats.
A Multi-tier Model and Filtering Approach to Detect Fake News Using Machine Learning Algorithms Chang Yu, Chiung; A Hamid, Isredza Rahmi; Abdullah, Zubaile; Kipli, Kuryati; Amnur, Hidra
JOIV : International Journal on Informatics Visualization Vol 8, No 2 (2024)
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.62527/joiv.8.2.2703

Abstract

Fake news trends have overgrown in our societies over the years through social media platforms. The goal of spreading fake news can easily mislead and manipulate the public’s opinion. Many previous researchers have proposed this domain using classification algorithms or deep learning techniques. However, machine learning algorithms still suffer from high margin error, which makes them unreliable as every algorithm uses a different way of prediction. Deep learning requires high computation power and a large dataset to operate the classification model. A filtering model with a consensus layer in a multi-tier model is introduced in this research paper. The multi-tier model filters the news label correctly predicted by the first two-tier layer. The consensus layer acts as the final decision when collision results occur in the first two-tier layer. The proposed model is applied to the WEKA software tool to test and evaluate the model from both datasets. Two sequences of classification models are used in this research paper: LR_DT_RF and LR_NB_AdaBoost. The best performance of sequence for both datasets is LR_DT_RF which yields 0.9892 F1-Score, 0.9895 Accuracy, and 0.9790 Matthews Correlation Coefficient (MCC) for ISOT Fake News Dataset, and 0.9913 F1-Score, 0.9853 Accuracy, and 0.9455 MCC for CHECKED Dataset. This research could give researchers an approach for fake news detection on different social platforms and feature-based
Entropy Based Method for Malicious File Detection Edzuan Zainodin, Muhammad; Zakaria, Zalmiyah; Hassan, Rohayanti; Abdullah, Zubaile
JOIV : International Journal on Informatics Visualization Vol 6, No 4 (2022)
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.30630/joiv.6.4.1265

Abstract

Ransomware is by no means a recent invention, having existed as far back as 1989, yet it still poses a real threat in the 21st century. Given the increasing number of computer users in recent years, this threat will only continue to grow, affecting more victims as well as increasing the losses incurred towards the people and organizations impacted in a successful attack. In most cases, the only remaining courses of action open to victims of such attacks were the following: either pay the ransom or lose their data. One commonly shared behavior by all crypto ransomware strains is that there will be attempts to encrypt the victims’ files at a certain point during the ransomware execution. This paper demonstrates a technique that can identify when these encrypted files are being generated and is independent of the strain of the ransomware. Previous research has highlighted the difficulty in differentiating between compressed and encrypted files using Shannon entropy, as both file types exhibit similar values. Among the experiments described in this study, one showed a unique characteristic for the Shannon entropy of encrypted file header fragments, which was used to differentiate between encrypted files and other high entropy files such as archives. The Shannon entropy of encrypted file header fragments has a unique characteristic in one of the tests discussed in this study. This property was used to distinguish encrypted files from other files with high entropy, such as archives. To overcome this drawback, this study proposed an approach for test case generation by enhancing the entropy-based threat tree model, which would improve malicious file identification. The file identification was enhanced by combining three entropy algorithms, and the test case was generated based on the threat tree model. This approach was then evaluated using accuracy measurements: True Positive, True Negative, False Positive, False Negative. A promising result is expected. This method solves the challenge of leveraging file entropy to distinguish compressed and archived files from ransomware-encrypted files in a timely manner.
Verification of Ph.D. Certificate using QR Code on Blockchain Ethereum Noorhizama, Nur Khairunnisa; Abdullah, Zubaile; Kasim, Shahreen; A Hamid, Isredza Rahmi; Mat Isa, Mohd Anuar
JOIV : International Journal on Informatics Visualization Vol 7, No 3 (2023)
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.30630/joiv.7.3.1584

Abstract

One of the major challenges the university faces is to provide real-time verification of their student's degree certification upon request by other parties. Conventional verification systems are typically costly, time-consuming and bureaucratic against certificate credential misconduct. In addition, the forgery of graduation degree certificates has become more efficient due to easy-to-use scanning, editing, and printing technologies. Therefore, this research proposes verifying Ph.D. certificates using QR codes on the Ethereum blockchain to address certificate verification challenges. Blockchain technology ensures tamper-proof and decentralized management of degree certificates as the certificates stored on the blockchain are replicated across the network. The issuance of certificates requires the use of the issuer's private key, thus preventing forgery. The system was developed using Solidity for the smart contract, PHP, HTML/CSS for the web-based implementation, and MetaMask for blockchain integration. User testing confirmed the successful implementation and functionality of the system. Users can add, update, and delete certificates, generate and scan QR codes, and receive instant verification feedback. The verification system effectively meets all requirements, providing a robust solution for validating Ph.D. certificates. Future research may focus on scalability and adoption, privacy and data protection, user experience, and integration with existing systems. Other researchers can optimize the verification system for widespread adoption and utilization by exploring these areas. This research contributes to securing and efficiently verifying academic certificates using QR codes on the Ethereum blockchain. Ultimately, this work advances the field of certificate verification and promotes trust in academic credentials.
A Microarray Data Pre-processing Method for Cancer Classification Hui, Tay Xin; Kasim, Shahreen; Md Fudzee, Mohd Farhan; Abdullah, Zubaile; Hassan, Rohayanti; Erianda, Aldo
JOIV : International Journal on Informatics Visualization Vol 6, No 4 (2022)
Publisher : Society of Visual Informatics

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.30630/joiv.6.4.1523

Abstract

The development of microarray technology has led to significant improvements and research in various fields. With the help of machine learning techniques and statistical methods, it is now possible to organize, analyze, and interpret large amounts of biological data to uncover significant patterns of interest. The exploitation of microarray data is of great challenge for many researchers. Raw gene expression data are usually vulnerable to missing values, noisy data, incomplete data, and inconsistent data. Hence, processing data before being applied for cancer classification is important. In order to extract the biological significance of microarray gene expression data, data pre-processing is a necessary step to obtain valuable information for further analysis and address important hypotheses. This study presents a detailed description of pre-processing data method for cancer classification. The proposed method consists of three phases: data cleaning, transformation, and filtering. The combination of GenePattern software tool and Rstudio was utilized to implement the proposed data pre-processing method. The proposed method was applied to six gene expression datasets: lung cancer dataset, stomach cancer dataset, liver cancer dataset, kidney cancer dataset, thyroid cancer dataset, and breast cancer dataset to demonstrate the feasibility of the proposed method for cancer classification. A comparison has been made to illustrate the differences between the dataset before and after data pre-processing.