Claim Missing Document
Check
Articles

Found 2 Documents
Search

Baby Cry Sound Detection: A Comparison of Mel Spectrogram Image on Convolutional Neural Network Models Junaidi, Ridha Fahmi; Faisal, Mohammad Reza; Farmadi, Andi; Herteno, Rudy; Nugrahadi, Dodon Turianto; Ngo, Luu Duc; Abapihi, Bahriddin
Journal of Electronics, Electromedical Engineering, and Medical Informatics Vol 6 No 4 (2024): October
Publisher : Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYA

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.35882/jeeemi.v6i4.465

Abstract

Baby cries contain patterns that indicate their needs, such as pain, hunger, discomfort, colic, or fatigue. This study explores the use of Convolutional Neural Network (CNN) architectures for classifying baby cries using Mel Spectrogram images. The primary objective of this research is to compare the effectiveness of various CNN architectures such as VGG-16, VGG-19, LeNet-5, AlexNet, ResNet-50, and ResNet-152 in detecting baby needs based on their cries. The datasets used include the Donate-a-Cry Corpus and Dunstan Baby Language. The results show that AlexNet achieved the best performance with an accuracy of 84.78% on the Donate-a-Cry Corpus dataset and 72.73% on the Dunstan Baby Language dataset. Other models like ResNet-50 and LeNet-5 also demonstrated good performance although their computational efficiency varied, while VGG-16 and VGG-19 exhibited lower performance. This research provides significant contributions to the understanding and application of CNN models for baby cry classification. Practical implications include the development of baby cry detection applications that can assist parents and healthcare provide.
Early Fusion of CNN Features for Multimodal Biometric Authentication from ECG and Fingerprint Using MLP, LSTM, GCN, and GAT Priyatama, Muhammad Abdhi; Nugrahadi, Dodon Turianto; Budiman, Irwan; Farmadi, Andi; Faisal, Mohammad Reza; Purnama, Bedy; Adi, Puput Dani Prasetyo; Ngo, Luu Duc
Jurnal Teknik Informatika (Jutif) Vol. 6 No. 6 (2025): JUTIF Volume 6, Number 6, Desember 2025
Publisher : Informatika, Universitas Jenderal Soedirman

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.52436/1.jutif.2025.6.6.5299

Abstract

Traditional authentication methods such as PINs and passwords remain vulnerable to theft and hacking, demanding more secure alternatives. Biometric approaches address these weaknesses, yet unimodal systems like fingerprints or facial recognition are still prone to spoofing and environmental disturbances. This study aims to enhance biometric reliability through a multimodal framework integrating electrocardiogram (ECG) signals and fingerprint images. Fingerprint features were extracted using three deep convolutional networks—VGG16, ResNet50, and DenseNet121—while ECG signals were segmented around the first R-peak to produce feature vectors of varying dimensions. Both modalities were fused at the feature level using early fusion and classified with four deep learning algorithms: Multilayer Perceptron (MLP), Long Short-Term Memory (LSTM), Graph Convolutional Network (GCN), and Graph Attention Network (GAT). Experimental results demonstrated that the combination of VGG16 + LSTM and ResNet50 + LSTM achieved the highest identification accuracy of 98.75 %, while DenseNet121 + MLP yielded comparable performance. MLP and LSTM consistently outperformed GCN and GAT, confirming the suitability of sequential and feed-forward models for fused feature embeddings. By employing R-peak-based ECG segmentation and CNN-driven fingerprint features, the proposed system significantly improves classification stability and robustness. This multimodal biometric design strengthens protection against spoofing and impersonation, providing a scalable and secure authentication solution for high-security applications such as digital payments, healthcare, and IoT devices.