cover
Contact Name
-
Contact Email
-
Phone
-
Journal Mail Official
-
Editorial Address
-
Location
Kota yogyakarta,
Daerah istimewa yogyakarta
INDONESIA
International Journal of Advances in Intelligent Informatics
ISSN : 24426571     EISSN : 25483161     DOI : 10.26555
Core Subject : Science,
International journal of advances in intelligent informatics (IJAIN) e-ISSN: 2442-6571 is a peer reviewed open-access journal published three times a year in English-language, provides scientists and engineers throughout the world for the exchange and dissemination of theoretical and practice-oriented papers dealing with advances in intelligent informatics. All the papers are refereed by two international reviewers, accepted papers will be available on line (free access), and no publication fee for authors.
Arjuna Subject : -
Articles 12 Documents
Search results for , issue "Vol 10, No 1 (2024): February 2024" : 12 Documents clear
Hybrid machine learning model based on feature decomposition and entropy optimization for higher accuracy flood forecasting Mohd Khairudin, Nazli; Mustapha, Norwati; Mohd Aris, Teh Noranis; Zolkepli, Maslina
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1130

Abstract

The advancement of machine learning model has widely been adopted to provide flood forecast. However, the model must deal with the challenges to determine the most important features to be used in in flood forecast with high-dimensional non-linear time series when involving data from various stations. Decomposition of time-series data such as empirical mode decomposition, ensemble empirical mode decomposition and discrete wavelet transform are widely used for optimization of input; however, they have been done for single dimension time-series data which are unable to determine relationships between data in high dimensional time series.  In this study, hybrid machine learning models are developed based on this feature decomposition to forecast the monthly water level using monthly rainfall data. Rainfall data from eight stations in Kelantan River Basin are used in the hybrid model. To effectively select the best rainfall data from the multi-stations that provide higher accuracy, these rainfall data are analyzed with entropy called Mutual Information that measure the uncertainty of random variables from various stations. Mutual Information act as optimization method helps the researcher to select the appropriate features to score higher accuracy of the model. The experimental evaluations proved that the hybrid machine learning model based on the feature decomposition and ranked by Mutual Information can increase the accuracy of water level forecasting.  This outcome will help the authorities in managing the risk of flood and helping people in the evacuation process as an early warning can be assigned and disseminate to the citizen.
A comparison of machine learning methods for knowledge extraction model in A LoRa-Based waste bin monitoring system Abidin, Aa Zezen Zaenal; Othman, Mohd Fairuz Iskandar; Hassan, Aslinda; Murdianingsih, Yuli; Suryadi, Usep Tatang; Siallagan, Timbo Faritchan
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1026

Abstract

Knowledge Extraction Model (KEM) is a system that extracts knowledge through an IoT-based smart waste bin emptying scheduling classification. Classification is a difficult problem and requires an efficient classification method. This research contributes in the form of the KEM system in the classification of scheduling for emptying waste bins with the best performance of the Machine Learning method. The research aims to compare the performance of Machine Learning methods in the form of Decision Tree, Naïve Bayes, K-Nearest Neighbor, Support Vector Machine, and Multi-Layer Perceptron, which will be recommended in the KEM system. Performance testing was performed on accuracy, recall, precision, F-Measure, and ROCS curves using the cross-validation method with ten observations. The experimental results show that the Decision Tree performs best for accuracy, recall, precision, and ROCS curve. In contrast, the K-NN method obtains the highest F-measure performance. KEM can be implemented to extract knowledge from data sets created in various other IoT-based systems.
Region-based convolutional neural networks for occluded person re-identification Islam, Atiqul; Tsun, Mark Tee Kit; Theng, Lau Bee; Chua, Caslon
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1125

Abstract

In a variety of applications, including intelligent surveillance systems, targeted tracking, and assistive human-following robots, the ability to accurately identify individuals even when they are partially obscured is imperative. Such Continuous person tracking is complicated by the close similarity between the appearance of people and target occlusions. This study addresses this significant challenge by proposing a two-step, detection-first approach that uses a region-based convolutional neural network (R-CNN) as the re-identification (re-ID)solution. The model is specifically trained to detect occluded persons at different levels of occlusion before forwarding the image for the re-ID process. Three occluded-specific datasets are selected to evaluate the model's effectiveness in detecting occluded people. There are 379 distinct people in total, and each has five images obstructed from different angles. A sample of the data is taken to simulate various environment settings, and new data points are generated with different degrees of occlusion to assess how well the model performs under varying levels of obstruction. The findings demonstrate that the proposed person re-ID model is reliable in most circumstances, correctly re-identifying at 74% (Rank-1) and 90% (Rank-5). Although there is a decrease in accuracy as the number of distinctive people in the dataset increases, this does not significantly impact the tracking performance in various applications, which are expected to recognize a single person or a small group of individuals. Future works will explore refining similarity matching algorithms by delving into robust image comparison techniques, thereby addressing the challenges presented by occlusions. A critical aspect is to assess the model under diverse lighting conditions and investigate scenarios with multiple individuals in a frame. It is also beneficial to exploit high-resolution datasets, such as DukeMTMC-reID, and integrate finer contextual details, like clothing or carried objects. These collective efforts are essential for optimizing the model’s efficacy in practical applications and advancing person re-ID technologies.
Emergency sign language recognition from variant of convolutional neural network (CNN) and long short term memory (LSTM) models As'ari, Muhammad Amir; Sufri, Nur Anis Jasmin; Qi, Guat Si
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1170

Abstract

Sign language is the primary communication tool used by the deaf community and people with speaking difficulties, especially during emergencies. Numerous deep learning models have been proposed to solve the sign language recognition problem. Recently. Bidirectional LSTM (BLSTM) has been proposed and used in replacement of Long Short-Term Memory (LSTM) as it may improve learning long-team dependencies as well as increase the accuracy of the model. However, there needs to be more comparison for the performance of LSTM and BLSTM in LRCN model architecture in sign language interpretation applications. Therefore, this study focused on the dense analysis of the LRCN model, including 1) training the CNN from scratch and 2) modeling with pre-trained CNN, VGG-19, and ResNet50. Other than that, the ConvLSTM model, a special variant of LSTM designed for video input, has also been modeled and compared with the LRCN in representing emergency sign language recognition. Within LRCN variants, the performance of a small CNN network was compared with pre-trained VGG-19 and ResNet50V2. A dataset of emergency Indian Sign Language with eight classes is used to train the models. The model with the best performance is the VGG-19 + LSTM model, with a testing accuracy of 96.39%. Small LRCN networks, which are 5 CNN subunits + LSTM and 4 CNN subunits + BLSTM, have 95.18% testing accuracy. This performance is on par with our best-proposed model, VGG + LSTM. By incorporating bidirectional LSTM (BLSTM) into deep learning models, the ability to understand long-term dependencies can be improved. This can enhance accuracy in reading sign language, leading to more effective communication during emergencies.
Analyzing computer vision models for detecting customers: a practical experience in a mexican retail Fernández Del Carpio, Alvaro
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1112

Abstract

Computer vision has become an important technology for obtaining meaningful data from visual content and providing valuable information for enhancing security controls, marketing, and logistic strategies in diverse industrial and business sectors. The retail sector constitutes an important part of the worldwide economy. Analyzing customer data and shopping behaviors has become essential to deliver the right products to customers, maximize profits, and increase competitiveness. In-person shopping is still a predominant form of retail despite the appearance of online retail outlets. As such, in-person retail is adopting computer vision models to monitor store products and customers. This research paper presents the development of a computer vision solution by Lytica Company to detect customers in Steren’s physical retail stores in Mexico. Current computer vision models such as SSD Mobilenet V2, YOLO-FastestV2, YOLOv5, and YOLOXn were analyzed to find the most accurate system according to the conditions and characteristics of the available devices. Some of the challenges addressed during the analysis of videos were obstruction and proximity of the customers, lighting conditions, position and distance of the camera concerning the customer when entering the store, image quality, and scalability of the process. Models were evaluated with the F1-score metric: 0.64 with YOLO FastestV2, 0.74 with SSD Mobilenetv2, 0.86 with YOLOv5n, 0.86 with YOLOv5xs, and 0.74 with YOLOXn. Although YOLOv5 achieved the best performance, YOLOXn presented the best balance between performance and FPS (frames per second) rate, considering the limited hardware and computing power conditions.
Imputation of missing microclimate data of coffee-pine agroforestry with machine learning Nurwarsito, Heru; Suprayogo, Didik; Sakti, Setyawan Purnomo; Prayogo, Cahyo; Yudistira, Novanto; Fauzi, Muhammad Rifqi; Oakley, Simon; Mahmudy, Wayan Firdaus
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1439

Abstract

This research presents a comprehensive analysis of various imputation methods for addressing missing microclimate data in the context of coffee-pine agroforestry land in UB Forest. Utilizing Big data and Machine learning methods, the research evaluates the effectiveness of imputation missing microclimate data with Interpolation, Shifted Interpolation, K-Nearest Neighbors (KNN), and Linear Regression methods across multiple time frames - 6 hours, daily, weekly, and monthly. The performance of these methods is meticulously assessed using four key evaluation metrics Mean Absolute Error (MAE), Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Mean Absolute Percentage Error (MAPE). The results indicate that Linear Regression consistently outperforms other methods across all time frames, demonstrating the lowest error rates in terms of MAE, MSE, RMSE, and MAPE. This finding underscores the robustness and precision of Linear Regression in handling the variability inherent in microclimate data within agroforestry systems. The research highlights the critical role of accurate data imputation in agroforestry research and points towards the potential of machine learning techniques in advancing environmental data analysis. The insights gained from this research contribute significantly to the field of environmental science, offering a reliable methodological approach for enhancing the accuracy of microclimate models in agroforestry, thereby facilitating informed decision-making for sustainable ecosystem management.
An automated learning method of semantic segmentation for train autonomous driving environment understanding Wang, Yang; Chen, Yihao; Yuan, Hao; Wu, Cheng
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1521

Abstract

One of the major reasons for the explosion of autonomous driving in recent years is the great development of computer vision. As one of the most fundamental and challenging problems in autonomous driving, environment understanding has been widely studied. It directly determines whether the entire in-vehicle system can effectively identify surrounding objects of vehicles and make correct path planning. Semantic segmentation is the most important means of environment understanding among the many image recognition algorithms used in autonomous driving. However, the success of semantic segmentation models is highly dependent on human expertise in data preparation and hyperparameter optimization, and the tedious process of training is repeated over and over for each new scene. Automated machine learning (AutoML) is a research area for this problem that aims to automate the development of end-to-end ML models. In this paper, we propose an automatic learning method for semantic segmentation based on reinforcement learning (RL), which can realize automatic selection of training data and guide automatic training of semantic segmentation. The results show that our scheme converges faster and has higher accuracy than researchers manually training semantic segmentation models, while requiring no human involvement.
Optimization of use case point through the use of metaheuristic algorithm in estimating software effort Ardiansyah, Ardiansyah; Zulfa, Mulki Indana; Tarmuji, Ali; Jabbar, Farisna Hamid
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1298

Abstract

Use Case Points estimation framework relies on the complexity weight parameters to estimate software development projects. However, due to the discontinue parameters, it lead to abrupt weight classification and results in inaccurate estimation. Several research studies have addressed these weaknesses by employing various approaches, including fuzzy logic, regression analysis, and optimization techniques. Nevertheless, the utilization of optimization techniques to determine use case weight parameter values has yet to be extensively explored, with the potential to enhance accuracy further. Motivated by this, the current research delves into various metaheuristic search-based algorithms, such as genetic algorithms, Firefly algorithms, Reptile search algorithms, Particle swarm optimization, and Grey Wolf optimizers. The experimental investigation was carried out using a Silhavy UCP estimation dataset, which contains 71 project data from three software houses and is publicly available. Furthermore, we compared the performance between models based on metaheuristic algorithms. The findings indicate that the performance of the Firefly algorithm outperforms the others based on five accuracy metrics: mean absolute error, mean balance relative error, mean inverted relative error, standardized accuracy, and effect size. This research could be useful for software project managers to leverage the practical implications of this study by utilizing the UCP estimation method, which is optimized using the Firefly algorithm.
Domain adaptation for driver's gaze mapping for different drivers and new environments Sonom-Ochir, Ulziibayar; Karungaru, Stephen; Terada, Kenji; Ayush, Altangerel
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1168

Abstract

Distracted driving is a leading cause of traffic accidents, and often arises from a lack of visual attention on the road. To enhance road safety, monitoring a driver's visual attention is crucial. Appearance-based gaze estimation using deep learning and Convolutional Neural Networks (CNN) has shown promising results, but it faces challenges when applied to different drivers and environments. In this paper, we propose a domain adaptation-based solution for gaze mapping, which aims to accurately estimate a driver's gaze in diverse drivers and new environments. Our method consists of three steps: pre-processing, facial feature extraction, and gaze region classification. We explore two strategies for input feature extraction, one utilizing the full appearance of the driver and environment and the other focusing on the driver's face. Through unsupervised domain adaptation, we align the feature distributions of the source and target domains using a conditional Generative Adversarial Network (GAN). We conduct experiments on the Driver Gaze Mapping (DGM) dataset and the Columbia Cave-DB dataset to evaluate the performance of our method. The results demonstrate that our proposed method reduces the gaze mapping error, achieves better performance on different drivers and camera positions, and outperforms existing methods. We achieved an average Strictly Correct Estimation Rate (SCER) accuracy of 81.38% and 93.53% and Loosely Correct Estimation Rate (LCER) accuracy of 96.69% and 98.9% for the two strategies, respectively, indicating the effectiveness of our approach in adapting to different domains and camera positions. Our study contributes to the advancement of gaze mapping techniques and provides insights for improving driver safety in various driving scenarios.
Self-supervised few-shot learning for real-time traffic sign classification Nguyen, Anh-Khoa Tho; Tran, Tin; Nguyen, Phuc Hong; Dinh, Vinh Quang
International Journal of Advances in Intelligent Informatics Vol 10, No 1 (2024): February 2024
Publisher : Universitas Ahmad Dahlan

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.26555/ijain.v10i1.1522

Abstract

Although supervised approaches for traffic sign classification have demonstrated excellent performance, they are limited to classifying several traffic signs defined in the training dataset. This prevents them from being applied to different domains, i.e., different countries. Herein, we propose a self-supervised approach for few-shot learning-based traffic sign classification. A center-awareness similarity network is designed for the traffic sign problem and trained using an optical flow dataset. Unlike existing supervised traffic sign classification methods, the proposed method does not depend on traffic sign categories defined by the training dataset. It applies to any traffic signs from different countries. We construct a Korean traffic sign classification (KTSC) dataset, including 6000 traffic sign samples and 59 categories. We evaluate the proposed method with baseline methods using the KTSC, German traffic sign, and Belgian traffic sign classification datasets. Experimental results show that the proposed method extends the ability of existing supervised methods and can classify any traffic sign, regardless of region/country dependence. Furthermore, the proposed approach significantly outperforms baseline methods for patch similarity. This approach provides a flexible and robust solution for classifying traffic signs, allowing for accurate categorization of every traffic sign, regardless of regional or national differences.

Page 1 of 2 | Total Record : 12