Please use this identifier to cite or link to this item: http://localhost:8081/jspui/handle/123456789/20345
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSharma, Kartik-
dc.date.accessioned2026-04-09T07:59:39Z-
dc.date.available2026-04-09T07:59:39Z-
dc.date.issued2022-05-
dc.identifier.urihttp://localhost:8081/jspui/handle/123456789/20345-
dc.guideSharma, Rakshaen_US
dc.description.abstractImage captioning in the general sense, is the process of generating a concise and clear description of an input image. For automatic image captioning there are some challenges in how to extract visual information from the provided image and how to transform this information into a proper meaningful text in a natural language. We can summarise the task of Automatic Image Caption generation as follows: Given an Image automatically generate a caption of the image describing the objects in the image and their relationships in Natural Language ❖ Detect Objects in the scene. ❖ Detect Relationship between the objects. ❖ Describe the information using Natural Language. All these phases are carried out sequentially to form meaningful captions from the image. First we need to detect objects in the scene so that their relationship can be understood, the number of objects and the depth of the relationship captured both create an impact on the end result i.e. the generated captions and depend on various factors such as the model chosen, type of attention etc. Although being able to automatically describe the contents of an image in a natural language can be a very challenging task, it also has a high pay-off as we will see ahead.en_US
dc.language.isoenen_US
dc.publisherIIT, Roorkeeen_US
dc.titleIMAGE CAPTIONING USING NEURAL NETWORKSen_US
dc.typeDissertationsen_US
Appears in Collections:MASTERS' THESES (CSE)

Files in This Item:
File Description SizeFormat 
20535035_Kartik Sharma.pdf4.1 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.