Claim Missing Document
Check
Articles

Found 3 Documents
Search
Journal : Proceeding of the Electrical Engineering Computer Science and Informatics

Dynamic Hand Gesture Recognition Using Temporal-Stream Convolutional Neural Networks Fladio Armandika; Esmeralda Contessa Djamal; Fikri Nugraha; Fatan Kasyidi
Proceeding of the Electrical Engineering Computer Science and Informatics Vol 7, No 1: EECSI 2020
Publisher : IAES Indonesia Section

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.11591/eecsi.v7.2048

Abstract

Movement recognition is a hot issue in machine learning. The gesture recognition is related to video processing, which gives problems in various aspects. Some of them are separating the image against the background firmly. This problem has consequences when there are incredibly different settings from the training data. The next challenge is the number of images processed at a time that forms motion. Previous studies have conducted experiments on the Deep Convolutional Neural Network architecture to detect actions on sequential model balancing each other on frames and motion between frames. The challenge of identifying objects in a temporal video image is the number of parameters needed to do a simple video classification so that the estimated motion of the object in each picture frame is needed. This paper proposed the classification of hand movement patterns with the Single Stream Temporal Convolutional Neural Networks approach. This model was robust against extreme non-training data, giving an accuracy of up to 81,7%. The model used a 50 layers ResNet architecture with recorded video training.
Semantic Classification of Scientific Sentence Pair Using Recurrent Neural Network Agung Besti; Ridwan Ilyas; Fatan Kasyidi; Esmeralda Contessa Djamal
Proceeding of the Electrical Engineering Computer Science and Informatics Vol 7, No 1: EECSI 2020
Publisher : IAES Indonesia Section

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.11591/eecsi.v7.2051

Abstract

One development of Natural Language Processing is the semantic classification of sentences and documents. The challenge is finding relationships between words and between documents through a computational model. The development of machine learning makes it possible to try out various possibilities that provide classification capabilities. This paper proposes the semantic classification of sentence pairs using Recurrent Neural Networks (RNN) and Long Short-Term Memory (LSTM). Each couple of sentences is turned into vectors using Word2Vec. Experiments carried out using CBOW and Skip-Gram to get the best combination. The results are obtained that word embedding using CBOW produces better than Skip-Gram, although it is still around 5%. However, CBOW slows slightly at the beginning of iteration but is stable towards convergence. Classification of all six classes, namely Equivalent, Similar, Specific, No Alignment, Related, and Opposite. As a result of the unbalanced data set, the retraining was conducted by eliminating a few classes member from the data set, thus providing an accuracy of 73% for non-training data. The results showed that the Adam model gave a faster convergence at the start of training compared to the SGD model, and AdaDelta, which was built, gave 75% better accuracy with an F1-Score of 67%.
Hand Movement Identification Using Single-Stream Spatial Convolutional Neural Networks Aldi Sidik Permana; Esmeralda Contessa Djamal; Fikri Nugraha; Fatan Kasyidi
Proceeding of the Electrical Engineering Computer Science and Informatics Vol 7, No 1: EECSI 2020
Publisher : IAES Indonesia Section

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.11591/eecsi.v7.2055

Abstract

Human-robot interaction can be through several ways, such as through device control, sounds, brain, and body, or hand gesture. There are two main issues: the ability to adapt to extreme settings and the number of frames processed concerning memory capabilities. Although it is necessary to be careful with the selection of the number of frames so as not to burden the memory, this paper proposed identifying hand gesture of video using Spatial Convolutional Neural Networks (CNN). The sequential image's spatial arrangement is extracted from the frames contained in the video so that each frame can be identified as part of one of the hand movements. The research used VGG16, as CNN architecture is concerned with the depth of learning where there are 13 layers of convolution and three layers of identification. Hand gestures can only be identified into four movements, namely 'right', 'left', 'grab', and 'phone'. Hand gesture identification on the video using Spatial CNN with an initial accuracy of 87.97%, then the second training increased to 98.05%. Accuracy was obtained after training using 5600 training data and 1120 test data, and the improvement occurred after manual noise reduction was performed.