Please use this identifier to cite or link to this item: http://artemis.cslab.ece.ntua.gr:8080/jspui/handle/123456789/19595
Full metadata record
DC FieldValueLanguage
dc.contributor.authorΣτάμου, Πηνελόπη-
dc.date.accessioned2025-05-03T16:01:13Z-
dc.date.available2025-05-03T16:01:13Z-
dc.date.issued2025-03-21-
dc.identifier.urihttp://artemis.cslab.ece.ntua.gr:8080/jspui/handle/123456789/19595-
dc.description.abstractVision-Language Models (VLMs) have demonstrated remarkable capabilities in complex visio-linguistic tasks. An extensive body of work has explored how prompting techniques and fine-tuning methods can be used to enhance their performance. However, modern multimodal LLMs still struggle with tasks that require complex reasoning, external knowledge, and human-aligned responses. In this work, we investigate the limitations of large-scale, multimodal models in handling open-ended tasks that demand external knowledge and commonsense reasoning. Focusing on the Stanford Image Paragraph Captioning and OK-VQA datasets, we find that although these models demonstrate substantial cognitive, linguistic, and reasoning abilities, their performance deteriorates when managing complex tasks simultaneously while adhering to specific response formats. Our analysis reveals that state-of-the-art multimodal models surpass existing datasets in paragraph generation but continue to face challenges in generating high-quality paragraphs. Similarly, they continue to struggle with knowledge-based, open-ended benchmarks such as OK-VQA. To boost their performance in the latter, we employ a collaborative framework comprising three models: the Scout, an LVLM that takes an image as input and describes it in a paragraph; the Analyser, an LLM that generates an initial answer to the question based on the image description; and the Resolver, an LLM that extracts and formats the final answer based on a set of predefined rules. Our framework yields improved performance over the single-agent baseline, indicating the effectiveness of a collaborative approach.en_US
dc.languageenen_US
dc.subjectLarge Language Modelsen_US
dc.subjectMultimodal Large Language Modelsen_US
dc.subjectMulti-Agent Systemsen_US
dc.subjectKnowledge-Based Visual Question Answeringen_US
dc.subjectImage Paragraph Captioningen_US
dc.subjectVision-Language Modelsen_US
dc.titleEnhancing Vision-Language Models: The Role of LLMs in Augmenting Performance and Reasoningen_US
dc.description.pages120en_US
dc.contributor.supervisorΒουλόδημος Αθανάσιοςen_US
dc.departmentΤομέας Τεχνολογίας Πληροφορικής και Υπολογιστώνen_US
Appears in Collections:Διπλωματικές Εργασίες - Theses

Files in This Item:
File Description SizeFormat 
Diploma Thesis Stamou Penelope.pdfDiploma Thesis Stamou Penelope23.32 MBAdobe PDFView/Open


Items in Artemis are protected by copyright, with all rights reserved, unless otherwise indicated.