Multimodal diffusion framework for collaborative text image audio generation and applications
Abstract This paper presents a novel framework for collaborative generation across text, image, and audio modalities using an enhanced diffusion model architecture. We introduce a Hierarchical Cross-modal Alignment Network that establishes unified representations while preserving modality-specific c...
Saved in:
| Main Authors: | Junhua Wang, Ouya Zhang, Yuan Jiang |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Nature Portfolio
2025-07-01
|
| Series: | Scientific Reports |
| Subjects: | |
| Online Access: | https://doi.org/10.1038/s41598-025-05794-4 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Semantics-aware human motion generation from audio instructions
by: Zi-An Wang, et al.
Published: (2025-06-01) -
Multi-Level Feature Dynamic Fusion Neural Radiance Fields for Audio-Driven Talking Head Generation
by: Wenchao Song, et al.
Published: (2025-01-01) -
Multimodal Alzheimer’s disease recognition from image, text and audio
by: Byounghwa Lee, et al.
Published: (2025-08-01) -
Hierarchical cross-modal attention and dual audio pathways for enhanced multimodal sentiment analysis
by: D. Vamsidhar, et al.
Published: (2025-07-01) -
Multimodal Music Genre Classification of Sotho-Tswana Musical Videos
by: Osondu E. Oguike, et al.
Published: (2025-01-01)