Realistic Speech-Driven Talking Video Generation with Personalized Pose
In this work, we propose a method to transform a speaker’s speech information into a target character’s talking video; the method could make the mouth shape synchronization, expression, and body posture more realistic in the synthesized speaker video. This is a challenging task because changes of mo...
Saved in:
| Main Authors: | Xu Zhang, Liguo Weng |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Wiley
2020-01-01
|
| Series: | Complexity |
| Online Access: | http://dx.doi.org/10.1155/2020/6629634 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
SPPNet: Single-Person Human Parsing and Pose Estimation in RGB Videos
by: Aditi Verma, et al.
Published: (2025-03-01) -
Directive Speech Act Performed by Male and Female Speakers’ in Ted Talk Language Learning Videos
by: Donni Husnan, et al.
Published: (2025-02-01) -
Benchmarking the First Realistic Dataset for Speech Separation
by: Rawad MELHEM, et al.
Published: (2025-07-01) -
AMT-Net: Adversarial Motion Transfer Network With Disentangled Shape and Pose for Realistic Image Animation
by: Nega Asebe Teka, et al.
Published: (2025-01-01) -
Cross-Domain Person Re-Identification Based on Multi-Branch Pose-Guided Occlusion Generation
by: Pengnan Liu, et al.
Published: (2025-01-01)