TBKIN: Threshold-based explicit selection for enhanced cross-modal semantic alignments.
Vision-language models aim to seamlessly integrate visual and linguistic information for multi-modal tasks, demanding precise semantic alignments between image-text pairs while minimizing the influence of irrelevant data. While existing methods leverage intra-modal and cross-modal knowledge to enhan...
Saved in:
| Main Authors: | Zihan Guo, Xiang Shen, Chongqing Chen |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Public Library of Science (PLoS)
2025-01-01
|
| Series: | PLoS ONE |
| Online Access: | https://doi.org/10.1371/journal.pone.0325543 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
AlignFusionNet: Efficient Cross-Modal Alignment and Fusion for 3D Semantic Occupancy Prediction
by: Ziyi Xu, et al.
Published: (2025-01-01) -
Semantic enhancement and cross-modal interaction fusion for sentiment analysis in social media.
by: Guangyu Mu, et al.
Published: (2025-01-01) -
Enhancing Word Embeddings for Improved Semantic Alignment
by: Julian Szymański, et al.
Published: (2024-12-01) -
Sequence, gaze, and modal semantics: modal verb selection in German permission inquiries
by: Zinken Jörg, et al.
Published: (2025-01-01) -
Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing
by: Tom Sherborne, et al.
Published: (2023-11-01)