Claim Missing Document
Check
Articles

Found 2 Documents
Search

Exploring Strategies for Optimizing Mobilenetv2 Performance in Classification Tasks Through Transfer Learning and Hyperparameter Tuning with A Local Dataset from Kigezi, Uganda. Turihohabwe, Jack; Ssembatya Richard; Wasswa William
The Indonesian Journal of Computer Science Vol. 14 No. 1 (2025): The Indonesian Journal of Computer Science (IJCS)
Publisher : AI Society & STMIK Indonesia

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.33022/ijcs.v14i1.4436

Abstract

Abstract Background Deep learning has proved to very vital in numerous applications in recent years. However, the development of a model may require access to datasets. Training models on datasets may impose numerous challenges in terms of computational constraints, making it inefficient for limited computational environments and in this study a local dataset from Kigezi Uganda will be used. The study will also explore the strategies of optimizing the MobilenetV2 through transfer learning and hyper-tuning. Main Objective: This study explored the strategies for optimizing MobileNetV2 performance in performing classification tasks through transfer learning, data augmentation, and hyper parameter tuning with local data from Kigezi, Uganda. A total of 2,415 images is the dataset used and 9,660 images were obtained after data augmentation. Methodology The methodology used is experimentation using transfer learning and hyper-tuning of the model. Results The model layer freezing. Freezing All Layers except Final Dense Layer: Training accuracy: 90%, Testing accuracy: 85%. The model was not flexible enough to adapt to the new dataset, Unfreezing Top 10 Layers: Training accuracy: 92%, Testing accuracy: 88%. Moderate improvement observed, but still underperforming. Unfreezing Top 20 Layers: Training accuracy: 95%, Testing accuracy: 91%. Significant improvement, suggesting that more layers need to be fine-tuned. Unfreezing Entire Network: Training accuracy: 98%, Testing accuracy: 96%. The model showed substantial improvement in learning task-specific features. Hyper tuning the Learning Rate. The optimal configuration was found by unfreezing the entire network, which allowed the model to fine-tune all layers, thus improving the model’s ability to generalize to the new dataset. Learning Rate Tuning: Learning rate is one of the most crucial hyper parameters. An extensive grid search was performed over the following values: 0.1, 0.01, 0.001, 0.0001, and 0.00001, Batch Size Tuning: Different batch sizes (16, 32, 64, and 128) were tested to determine the most efficient size for gradient updates, Optimizer Selection: Various optimizers were tested, including SGD, RMSprop, and Adam. The Adam optimizer was selected for its adaptive learning rate capabilities. Epochs and Early Stopping: The number of epochs was tuned along with early stopping criteria to prevent overfitting. Epochs were tested in the range of 10 to 100 with a patience of 5 for early stopping The results of the learning rate 0.1: Training accuracy: 60%, Testing accuracy: 55%. The model was unable to converge 0.01: Training accuracy: 80%, Testing accuracy: 75%. Improved but still underperforming. 0.001: Training accuracy: 90%, Testing accuracy: 88%. Further improvement, but overfitting observed. 0.0001: Training accuracy: 99%, Testing accuracy: 98%. Optimal performance achieved.0.00001: Training accuracy: 95%, Testing accuracy: 92%. Learning was too slow. Hyper-tuning using the batch-size: 16: Training accuracy: 97%, Testing accuracy: 94%. Good performance but higher computational cost32: Training accuracy: 99%, Testing accuracy: 98%. Optimal balance between performance and efficiency, 64: Training accuracy: 95%, Testing accuracy: 93%. Slightly reduced performance, 128: Training accuracy: 90%, Testing accuracy: 87%. The model struggled with larger batch sizes. Hyper-tuning using by different optimizers SGD: Training accuracy: 85%, Testing accuracy: 80%. Slower convergence. RMSprop: Training accuracy: 92%, Testing accuracy: 88%. Moderate performance. Adam: Training accuracy: 99%, Testing accuracy: 98%. Best performance due to adaptive learning rate. The final customized model, after applying transfer learning and extensive hyper parameter tuning, achieved outstanding results: Training Accuracy: 99%Testing Accuracy: 98%,Training Loss: 0.02,Testing Loss: 0.04.
A Determination of Sample Size for Plant Leaves in Deep Learning Models for Predicting Late Blight in Irish Potatoes: An experimentation methodology in Kigezi –Uganda Turihohabwe, Jack
The Indonesian Journal of Computer Science Vol. 14 No. 2 (2025): The Indonesian Journal of Computer Science
Publisher : AI Society & STMIK Indonesia

Show Abstract | Download Original | Original Source | Check in Google Scholar | DOI: 10.33022/ijcs.v14i2.4647

Abstract

Determining the sample size of Deep learning models still remains a challenges in the Artificial Intelligence world. This is because most of the developers of deep learning models utilize available data collected from public datasets sites such PlantVillage or Kaggle. This study proposes using the acreage method putting into consideration of the machine learning dataset condition. The main objective of this research is to experiment the methods that can be used to determine the appropriate sample size for a Deep learning model. This study used the experimental and statistical methodologies and incorporated the boundaries of the Machine learning condition. The average sample estimation of the measurements in the piece of land (plot) was (1x4X10) cm. The measurement of the leaves was 3.5-5cm in length and 1.5-3 cm in width. The experiments were done between (2:00-4:00) am to have a good lighting condition. The optimal leaning rate of the deep learning architectures involved in the study used a learning rate of 0.0001. The study covered an acreage of 28000.25 acres and the Dataset 2145 Irish potato leaves was obtained and got 9,660 images after augmentation. This was purposively collected from ten sub-counties due to time and financial constraints in this study. This study proposed a methodology for obtaining the sample size using the acreage methodology and purposive sampling and there use the Machine learning condition for  sample sizes  for creation of deep learning models from potato leaf images targeted at preventing late blight based on leaf images. Future research may extend this study to further more validate the acreage methodology putting into account the Machine learning condition and also developing the Deep learning condition.