Science in Information Technology Letters
Vol 1, No 1: May 2020

Automated image captioning with deep neural networks

Abdullah Ahmad Zarir (Department of Computer Science, International Islamic University Malaysia)
Saad Bashar (Department of Computer Science, International Islamic University Malaysia)
Amelia Ritahani Ismail (Department of Computer Science, International Islamic University Malaysia)



Article Info

Publish Date
27 Apr 2020

Abstract

Generating natural language descriptions of the content of an image automatically is a complex task. Though it comes naturally to humans, it is not the same when making a machine do the same. But undoubtedly, achieving this feature would remarkably change how machines interact with us. Recent advancement in object recognition from images has led to the model of captioning images based on the relation between the objects in it. In this research project, we are demonstrating the latest technology and algorithms for automated caption generation of images using deep neural networks. This model of generating a caption follows an encoder-decoder strategy inspired by the language-translation model based on Recurrent Neural Networks (RNN). The language translation model uses RNN for both encoding and decoding, whereas this model uses a Convolutional Neural Networks (CNN) for encoding and an RNN for decoding. This combination of neural networks is more suited for generating a caption from an image. The model takes in an image as input and produces an ordered sequence of words, which is the caption.

Copyrights © 2020






Journal Info

Abbrev

sitech

Publisher

Subject

Computer Science & IT

Description

Science in Information Technology Letters (SITech) aims to keep abreast of the current development and innovation in the area of Science in Information Technology as well as providing an engaging platform for scientists and engineers throughout the world to share research results in related ...