Learning More May Not Be Better: Knowledge Transferability in Vision-and-Language Tasks
Is learning more knowledge always better for vision-and-language models? In this paper, we study knowledge transferability in multi-modal tasks. The current tendency in machine learning is to assume that by joining multiple datasets from different tasks, their overall performance improves. However,...
Saved in:
| Main Authors: | Tianwei Chen, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima, Hajime Nagahara |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
MDPI AG
2024-11-01
|
| Series: | Journal of Imaging |
| Subjects: | |
| Online Access: | https://www.mdpi.com/2313-433X/10/12/300 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Enhancing Deepfake Detection Through Quantum Transfer Learning and Class-Attention Vision Transformer Architecture
by: Bekir Eray Katı, et al.
Published: (2025-01-01) -
Dynamic Identification of Bridges Using Multiple Synchronized Cameras and Computer Vision
by: Tommaso Panigati, et al.
Published: (2025-02-01) -
Integration of Multi-Modal Sensors and Robot Arm Vision for Monitoring and Assisting Elderly Activities
by: Suwatri Jura, et al.
Published: (2025-03-01) -
Transfer Learning-Based Ensemble of CNNs and Vision Transformers for Accurate Melanoma Diagnosis and Image Retrieval
by: Murat Sarıateş, et al.
Published: (2025-07-01) -
Computer Vision and Transfer Learning for Grading of Egyptian Cotton Fibres
by: Ahmed Rady, et al.
Published: (2025-04-01)