Claim Missing Document
Check
Articles

Found 4 Documents
Search
Journal : Building of Informatics, Technology and Science

Eksperimen Pengujian Optimizer dan Fungsi Aktivasi Pada Code Clone Detection dengan Pemanfaatan Deep Neural Network (DNN) Errissya Rasywir; Yovi Pratama; Fachruddin Fachruddin
Building of Informatics, Technology and Science (BITS) Vol 4 No 2 (2022): September 2022
Publisher : Forum Kerjasama Pendidikan Tinggi

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47065/bits.v4i2.1776

Abstract

The problem of similarity (similarity) of program code can be one solution to the plagiarism detection approach. Plagiarism raises a form of action and consequences of plagiarism itself if the source used is not open source. Plagiarism is an act of deception of the work of others without the knowledge of the original author, which violates a Copyright and Moral Rights. With the increasing amount of data and data complexity, deep learning provides solutions for predictive analytics, with increased processing capabilities and optimal processor utilization. Deep learning shows success and improves the classification model in this field. On the other hand, clone detection code with massive, varied and high-speed data volumes requires feature extraction. With the potential of deep learning to extract better features, deep learning techniques are suitable for code clone detection. For this reason, it is necessary to develop a clone detection code that can process data from a programming language by utilizing deep learning. Based on the results of experiments conducted on 100 PHP program code data files, experimented with several types of activation function and optimizer methods. The average value of the resulting accuracy is good. With a variety of activation functions that we use such as Relu, Linear, Sigmoid, Softmax, Tanh, Elu, Selu, Softplus, Softsign, hard, and sigmoid, as well as variations of the optimizer used are Adagrad, RMSProp, SGD, Adadelta, Adam, Adamax and Nadam , the best attribute selection is in the Selu function and the RMSProp optimizer. The number of epochs used is 1000, the number of neurons per layer is 500 and the best number of hidden layers is 10, the average accuracy is 0.900
Penerapan Algoritma K-Means clustering Untuk Mengelompokkan Provinsi Berdasarkan Banyaknya Desa/Kelurahan Dengan Upaya Antisipasi/Mitigasi Bencana Alam Pratama, Yovi; Hendrawan, Hendrawan; Rasywir, Errissya; Carenina, Babel Tio; Anggraini, Dila Riski
Building of Informatics, Technology and Science (BITS) Vol 4 No 3 (2022): December 2022
Publisher : Forum Kerjasama Pendidikan Tinggi

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47065/bits.v4i3.2549

Abstract

Natural disasters are one of the natural phenomena that threaten human survival. The negative impacts can be in the form of material or non-material losses. However, with the ability to recognize the early symptoms of a disaster, humans can prepare themselves for disaster. Application of the K-Means clustering Algorithm in Grouping Provinces Based on the Number of Villages / Sub-districts with Anticipation / Mitigation Efforts for Natural Disasters Using the WEKA Application. The data sources for this research were collected based on documents describing the Number of Villages/ Urban According to Natural Disaster Anticipation/Mitigation Efforts produced by the National Statistics Agency. The data used in this study is provincial data which consists of 34 provinces. There are 4 variables used, namely Natural Disaster Early Warning System, Tsunami Early Warning System, Safety Equipment, Evacuation Path. The data will be processed by clustering in 2 clusters, namely clusters with high anticipation/mitigation levels and low anticipation/mitigation levels. The results obtained from the assessment process are that there are 5 (14.71 %) provinces with a high level of anticipation/mitigation and 29 (85.29%) other provinces including a low level of anticipation/mitigation. This can be an input for the government to pay more attention to the Village/Kelurahan based on the clusters that have been carried out
Eksperimen Layer Pooling menggunakan Standar Deviasi untuk Klasifikasi Dataset Citra Wajah dengan Metode CNN Pratama, Yovi; Rasywir, Errissya; Fachruddin, Fachruddin; Kisbianty, Desi; Irawan, Beni
Building of Informatics, Technology and Science (BITS) Vol 5 No 1 (2023): June 2023
Publisher : Forum Kerjasama Pendidikan Tinggi

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47065/bits.v5i1.3604

Abstract

Deep Learning, especially the Convolutional Neural Network (CNN) has proven to be reliable in processing data from various programming language platforms by utilizing deep learning. In this study, we modified it by calculating the statistical variance. The modifications made are replacing calculations on the Pooling Layer which generally use two formulas, namely max pooling and average pooling. We use the standard deviation to change the reduced image intensity value. With the research experiments built, it is expected to be able to perform facial recognition as an indicator for testing modifications. The Layer Pooling experiment uses the Standard Deviation for Classifying Face Image Datasets with the CNN Method, including the type of dataset used is the Aberdeen dataset https://pics.stir.ac.uk/2D_face_sets.htm. From the results of the experiments conducted, it was found that the highest value was using the Elu activation function and the Adagrad optimizer worth 77.844% for max pooling and 79.844% for pooling with a standard deviation. The Cellu activation function and the RMSprop optimizer are 77.986% for max pooling and 75.986% for pooling with a standard deviation. The highest score with the Softplus activation function and the Sgd optimizer is 77.844% for max pooling usage and 76.344% for pooling with standard deviation. The Tanh activation function and the Adadelta optimizer are 87.844% for max pooling and 85.844% for pooling with a standard deviation. The Elu activation function and the Adam optimizer are 87.853% for the use of max pooling and 85.285% for pooling with a standard deviation. By using the Elu activation function and the Adamax optimizer, the value is 87.842% for max pooling and 86.242% for pooling with a standard deviation. The highest score is using the Elu activation function and the Nadam optimizer with a value of 87.845% for max pooling usage and 86.345% for using standard deviation calculations as pixel pooling. From all experiments it was stated that the use of pooling with the highest value technique or max pooling still gave a better value than using the standard deviation calculation with the best tuning results using the Elu activation function and Adam's Optimiser, which was 87.853%.
Reduksi False Positive Pada Klasifikasi Job Placement dengan Hybrid Random Forest dan Auto Encoder Pahlevi, M. Riza; Rasywir, Errissya; Pratama, Yovi; Istoningtyas, Marrylinteri; Fachruddin, Fachruddin; Yaasin, Muhammad
Building of Informatics, Technology and Science (BITS) Vol 5 No 4 (2024): March 2024
Publisher : Forum Kerjasama Pendidikan Tinggi

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.47065/bits.v5i4.4864

Abstract

The False Positive (FP) interpretation shows a negative prediction result and is a type 1 error answer with an incorrect positive prediction result. Based on this, we try to reduce type 1 errors to increase the accuracy value of the classification results. A low FP rate is critical for the use of Computer Aided Detection (CAD) systems. In this research proposal, to reduce FP, we use a Random Forest (RF) evaluation result design which will be reinterpreted by the Auto Encoder (AE) algorithm. The RF algorithm was chosen because it is a type of ensemble learning that can optimize accuracy in parallel. RF was chosen because it performs bagging on all Decision Tree (DT) outputs used. To suppress TP reduction more strongly, we use the Auto Encoder (AE) algorithm to reprocess the class bagging results from RF into input in the AE layer. AE uses reconstruction errors, which in this case is Job Placement classification. From the test results, it was found that combining the use of a random forest using C4.5 as a decision tree with an Autoencoder can increase accuracy in the Job Placement Classification task by a difference of 0.004652 better than without combining it with an autoencoder. Apart from that, in testing using a combination of RF and AE, fewer False Positive (FP) values ​​were produced, namely 11 items in the Cross Validation-5 (CV-5) Test, then 13 items in the Cross Validation-10 (CV-10) test and in testing split training data of 60%, the FP was only 12. This value is less than the false positives produced by testing without Autoencoder, namely 12 items on CV-5, 15 items on CV-10, and 13 on split training data
Co-Authors Abdul Haris Abdul Harris Abdurrahman Abidin, Dodo Zaenal Abrani, Sauti Ade Saputra Agus Siswanto Akwan Sunoto Anggraini, Dila Riski Anita Anita Nurjanah Annisa putri Anton Prayitno Arya Atmanegara Aryani, Lies asih asmarani Athalina, Ghita Bayu saputra Beni Irawan Betantiyo Prayatna Borroek, Maria Rosario Briyan Chairullah Candra Adi Rahmat Carenina, Babel Tio Clara Zuliani Syahputri Defrin Azrian Desi Kisbianty, Desi Despita Meisak desy ayu ramadhanty Dimas Pratama Dodo Zaenal Abidin Dwi Rosa Indah Elsa Charolina L Siantar Evan Albert Fachruddin Fachruddin Fachruddin Fachruddin Fachruddin Fachruddin Fachruddin Fachruddin, Fachruddin farchan akbar Feranika, Ayu Fernando Fernando fiqri ansyah Fradea Novi Ramadhayanti GILLIANI, WENNY Hani Prastiwi Hartiwi, Yessi Hendrawan Hendrawan Hendrawan Hendrawan Hendrawan Hendrawan Hilda Permatasari Hussaein, Ahmad Ilham Adriansyah Ilham Fahrozi ilham permana Imelda Yose Iqbal Pradibya Irawan Irawan Irawan Irawan Irawan, Beni Istoningtyas, Marrylinteri Jasmir Jasmir Jeny Pricilia Johari, Riyan Jopi Mariyanto khalil gibran ahmad Kholil Ikhsan Lazuardi Yudha Pradana Li Sensia Rahmawati Lies Aryani Luthfi Rifky M.Rizky Wijaya Macharani Raschintasofi Maliyatul Khasanah Maria Rosario Borroek Marrylinteri Istoningtyas Marrylinteri Istoningtyas Marrylinteri Istoningtyas Mayang Ruza Mgs Afriyan Firdaus Migi Sulistiono Muhammad David Adrilyan Muhammad Diemas Mahendra Muhammad Ismail Muhammad Ismail Muhammad Riza Pahlevi Muhammad Satria Mubin Muhammad Wahyu Prayogi Mulyadi Mulyadi Mumtaz Ilham S Mumtaz Ilham Syafatullah Muttaqin Nabila Khumairo Najmul Laila Nanda Ghina Nasrul Ahlunaza Nasutioni, Wahyudi Nilu Widyawati Nungky Septia Kurnicova Nur Aini Nur Azmi Nurhadi Nurhadi Nurul Aulia OPHELIA, CHANDY Pahlevi, M. Riza Pahlevi, M.Riza Pareza Alam Jusia Pareza Alam Jusia Pareza Alam Jusia, Pareza Alam Putri Ratna Sari, Putri Ratna Rani Oktavia, Feby Renita Syafitri Reza Pahlevi Rio Ferdinand ROBY SETIAWAN Rofi'i, Imam Rohaini, Eni Rosario B, Maria Rosario, Maria Rts CiptaNingsi Rudolf Sinaga Sandi Pramadi Saparudin, Saparudin Satria Oldie Versileno Sri Wahyuni Nainggolan Sulistia Ramadhani Suyanti Tasya Basalia Sihombing Tedy Hardiyanto Tondy Maulana Tambunan Verwin Juniansyah virginia casanova andiko andiko Wahid Hasyim Yaasin, Muhammad Yessi Hartiwi Yessi Hartiwi Yoga Rizki Yovi Pratama Yuga Pramudya Zahlan Nugraha