PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.

Bind-Your-Avatar: Multi-Talking-Character Video Generation with Dynamic 3D-mask-based Embedding Router

arXivSource

Yubo Huang, Weiqiang Wang, Sirui Zhao, Tong Xu, Lin Liu, Enhong Chen

cs.CV
|
Jun 24, 2025
8 views

One-line Summary

Bind-Your-Avatar introduces a novel framework for generating videos with multiple talking characters in the same scene, using a dynamic 3D-mask embedding router to control audio-to-character correspondence and a new dataset for training and benchmarking.

Plain-language Overview

This research introduces a new approach to creating videos where multiple characters talk together in the same scene. Traditional methods often only handle one character at a time or separate conversations between two characters. The authors developed a system called Bind-Your-Avatar, which uses advanced techniques to ensure each character's speech matches their audio correctly. They also created a new dataset specifically for training these multi-character videos, allowing for better performance compared to existing methods.

Technical Details