Prompting Large Language Models with Knowledge-Injection for Knowledge-Based Visual Question Answering
Previous works employ the Large Language Model (LLM) like GPT-3 for knowledge-based Visual Question Answering (VQA). We argue that the inferential capacity of LLM can be enhanced through knowledge injection. Although methods that utilize knowledge graphs to enhance LLM have been explored in various...
Saved in:
Main Authors: | Zhongjian Hu, Peng Yang, Fengyuan Liu, Yuan Meng, Xingyu Liu |
---|---|
Format: | Article |
Language: | English |
Published: |
Tsinghua University Press
2024-09-01
|
Series: | Big Data Mining and Analytics |
Subjects: | |
Online Access: | https://www.sciopen.com/article/10.26599/BDMA.2024.9020026 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Knowledge Graphs as a source of trust for LLM-powered enterprise question answering
by: Juan Sequeda, et al.
Published: (2025-05-01) -
Assessing the performance of zero-shot visual question answering in multimodal large language models for 12-lead ECG image interpretation
by: Tomohisa Seki, et al.
Published: (2025-02-01) -
Visual Question Answering in Robotic Surgery: A Comprehensive Review
by: Di Ding, et al.
Published: (2025-01-01) -
Answer Distillation Network With Bi-Text-Image Attention for Medical Visual Question Answering
by: Hongfang Gong, et al.
Published: (2025-01-01) -
Enhancing students’ participation through question and answer on SMAN 2 Sungai Kakap Kubu Raya
by: Clarry Sada, et al.
Published: (2024-02-01)