Skip to content

Automatically captioning a unseen image using encoding with CNN and decoding with RNN.

Notifications You must be signed in to change notification settings

soheil-mp/Automatic-Image-Captioning

Repository files navigation

Automatic Image Captioning

Introduction

Image Captioning is the process of automatically captioning a unseen image. It uses both Natural Language Processing and Computer Vision to generate the captions. let's take a look at sample pictures which the captions has been generated:




Network Topology

Encoder:
We use Convolutional Neural Network(CNN) as our encoder. The image is given to CNN to extract the relavant features. The last hidden state in CNN is connected to Decoder The encoder that we provide to you uses the pre-trained ResNet-50 architecture (with the final fully-connected layer removed) to extract features from a batch of pre-processed images. The output is then flattened to a vector, before being passed through a Linear layer to transform the feature vector to have the same size as the word embedding.

Decoder:
We use Recurrent Neural Network(RNN) as our encoder which it takes the features from encoder and procuce a sectence for it.

Dataset

The Microsoft Common Objects in COntext (MS COCO) dataset is a large-scale dataset for scene understanding. The dataset is commonly used to train and benchmark object detection, segmentation, and captioning algorithms.

You can read more about the dataset on the website or in the research paper.

About

Automatically captioning a unseen image using encoding with CNN and decoding with RNN.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published