Exploring latent weight factors and global information for food-oriented cross-modal retrieval
Food-oriented cross-modal retrieval aims to retrieve relevant recipes given food images or vice versa. The modality semantic gap between recipes and food images (text and image modalities) is the main challenge. Though several studies are introduced to bridge this gap, they still suffer from two maj...
Saved in:
| Main Authors: | Wenyu Zhao, Dong Zhou, Buqing Cao, Wei Liang, Nitin Sukhija |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Taylor & Francis Group
2023-12-01
|
| Series: | Connection Science |
| Subjects: | |
| Online Access: | http://dx.doi.org/10.1080/09540091.2023.2233714 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Cross modal recipe retrieval with fine grained modal interaction
by: Fan Zhao, et al.
Published: (2025-02-01) -
Pic2Plate: A Vision-Language and Retrieval-Augmented Framework for Personalized Recipe Recommendations
by: Yosua Setyawan Soekamto, et al.
Published: (2025-01-01) -
Strong and Weak Prompt Engineering for Remote Sensing Image-Text Cross-Modal Retrieval
by: Tianci Sun, et al.
Published: (2025-01-01) -
DCLMA: Deep correlation learning with multi-modal attention for visual-audio retrieval
by: Jiwei Zhang, et al.
Published: (2025-09-01) -
DI-VTR: Dual inter-modal interaction model for video-text retrieval
by: Jie Guo, et al.
Published: (2024-09-01)