Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Need Advices on Best Parameters for JoyVasa in Video-to-Video Lip-Sync #110

Open
tonyabracadabra opened this issue Dec 16, 2024 · 1 comment

Comments

@tonyabracadabra
Copy link

I noticed the integration with JoyVasa—great to see advancements in this space! I'm particularly curious about how it can be leveraged for video-to-video lip-sync (focusing solely on lip movements, without eye or head motion). Specifically, I'm wondering what parameters work best to make it a viable replacement for older algorithms like Wav2Lip.

I've experimented with some parameter settings myself and read through discussions in the LivePortrait repo. The community seems to have diverging views—some advocate for relative motion, while others favor absolute motion.

Should I consider retargeting the lips for better accuracy? Additionally, what parameter configuration has empirically been shown to work best in your experience? Any insights would be greatly appreciated.

@warmshao
Copy link
Owner

I noticed the integration with JoyVasa—great to see advancements in this space! I'm particularly curious about how it can be leveraged for video-to-video lip-sync (focusing solely on lip movements, without eye or head motion). Specifically, I'm wondering what parameters work best to make it a viable replacement for older algorithms like Wav2Lip.

I've experimented with some parameter settings myself and read through discussions in the LivePortrait repo. The community seems to have diverging views—some advocate for relative motion, while others favor absolute motion.

Should I consider retargeting the lips for better accuracy? Additionally, what parameter configuration has empirically been shown to work best in your experience? Any insights would be greatly appreciated.

Currently, it defaults to driving with normalize_lip. How else would you like to retarget the lips? For me, relative motion combined with lip region animation works best.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants