Please use this identifier to cite or link to this item: http://drsr.daiict.ac.in//handle/123456789/1000
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorKhare, Manish
dc.contributor.authorPatel, Abhikumar
dc.contributor.otherKumar, Ahlad
dc.date.accessioned2022-05-06T05:46:40Z
dc.date.available2023-02-19T05:46:40Z
dc.date.issued2021
dc.identifier.citationPatel, Abhikumar (2021). Image Captioning Using Visual And Semantic Attention Mechanism. Dhirubhai Ambani Institute of Information and Communication Technology. viii, 40 p. (Acc.No: T00940)
dc.identifier.urihttp://drsr.daiict.ac.in//handle/123456789/1000
dc.description.abstractImage captioning is a method of generating captions/descriptions for the image. Image captioning have many applications in various fields like image indexing for content based image retrieval, Self-driving car, for visually impaired persons, in smart surveillance system and many more. It connects two major research communities of computer vision and natural language processing. The main challenges in image captioning are to recognize the important objects, their attributes, and their visual relationships of objects within an image, then it also needs to generate syntactically and semantically correct sentences. Currently, most of the architectures for image captioning are based on the encoder-decoder model, in which the image is first encoded using CNN to get an abstract version of the image then it is decoded using RNN to get proper caption for the image. So finally I have selected one base paper which was based on visual attention on the image to attend the most appropriate region of the image while generating each word for the caption. But they have miss one important factor while generating the caption for the image which was visual relationships between the objects present in the image. So I have decided to add one relationship detector module to that model to consider the relationships between objects. After combining this module with existing show-attend and tell model we get the caption for the image which consider the relationships between object, which ultimately enhance the quality of the caption for the image. I have performed experiments on various publicly available standard datasets like Flickr8k dataset, Flickr30k dataset and MSCOCO dataset.
dc.publisherDhirubhai Ambani Institute of Information and Communication Technology
dc.subjectEncoder
dc.subjectDecoder
dc.subjectConvolutional Neural Network
dc.subjectRecurrent Neural Network
dc.subjectVisual relationship detector
dc.subjectattention mechanism
dc.classification.ddc006.3 PAT
dc.titleImage Captioning Using Visual And Semantic Attention Mechanism
dc.typeDissertation
dc.degreeM. Tech
dc.student.id201911012
dc.accession.numberT00940
Appears in Collections:M Tech Dissertations

Files in This Item:
File Description SizeFormat 
201911012_Final MTT - Manish Khare.pdf
  Restricted Access
12.05 MBAdobe PDFView/Open Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.