Machine translation for low-resource languages remains a significant challenge due to the lack of parallel corpora and optimized model configurations. This study developed and optimized a Seq2Seq Long Short-Term Memory (LSTM) model for Tegalan-to-Indonesian translation. A manually curated parallel corpus was constructed to train and evaluate the model. Various hyperparameter configurations were systematically tested, with the best-performing model achieving a BLEU score of 11.7381 using a dropout rate of 0.5, batch size of 64, learning rate of 0.01, and 70 training epochs. The results demonstrated that higher dropout rates, smaller batch sizes, and longer training durations enhanced model generalization and translation accuracy. The optimized model was deployed into a web-based application using Streamlit, ensuring accessibility for real-time translation. The findings highlighted the importance of hyperparameter tuning in neural machine translation for low-resource languages. Future research should explore Transformer-based architectures, larger datasets, and reinforcement learning techniques to further enhance translation quality and generalization.
Copyrights © 2025