Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

whether the method proposed in this paper supports explicit key-point to drive source image? #83

Open
zsf23 opened this issue Jul 10, 2024 · 6 comments

Comments

@zsf23
Copy link

zsf23 commented Jul 10, 2024

since implicity key-points are used in this paper, what I'm concerned about is that explicit key-point, for example 106 2D landmark, can be use to drive source image?

@cleardusk
Copy link
Member

Yeap, I think it supports. But 2D landmarks are ambiguous, especially with large poses.

@cleardusk
Copy link
Member

Many diffusion-based methods use 2D landmarks or 3D-to-2D projected landmarks as the condition to animate.

@johndpope
Copy link

fyi - i recreated Samsung AI - megaportraits https://github.com/johndpope/MegaPortrait-hack -
it has fast warping / doesn't use keypoints
johndpope/MegaPortrait-hack#36

Microsoft mention in their VASA paper - they use this ORIGINAL resnet50 implementation (as opposed to the more recent MetaPortrait) johndpope/MegaPortrait-hack#16
I'm hoping EmoPortraits code drops this month - which will clear up some things.

@zsf23
Copy link
Author

zsf23 commented Jul 22, 2024

Many diffusion-based methods use 2D landmarks or 3D-to-2D projected landmarks as the condition to animate.

Thanks for your answer. I also get another questiones about formula (3) that mentioned in paper. 1) what are the indexes of 2d landmarks and implicit keypoints in formula (2). 2)the 2d landmarks are extracted from source image、drive image, or both of them?

@zzzweakman
Copy link
Collaborator

Many diffusion-based methods use 2D landmarks or 3D-to-2D projected landmarks as the condition to animate.

Thanks for your answer. I also get another questiones about formula (2) that mentioned in paper. 1) what are the indexes of 2d landmarks and implicit keypoints in formula (2). 2)the 2d landmarks are extracted from source image、drive image, or both of them?

In formula (2), the implicit keypoints are all 3D; there are no 2D landmarks included. Here, x_s and x_d represent the source and driving 3D implicit keypoints, respectively, while x_c,s denotes the canonical keypoints of the source image.

@zsf23
Copy link
Author

zsf23 commented Jul 29, 2024

Many diffusion-based methods use 2D landmarks or 3D-to-2D projected landmarks as the condition to animate.

Thanks for your answer. I also get another questiones about formula (2) that mentioned in paper. 1) what are the indexes of 2d landmarks and implicit keypoints in formula (2). 2)the 2d landmarks are extracted from source image、drive image, or both of them?

In formula (2), the implicit keypoints are all 3D; there are no 2D landmarks included. Here, x_s and x_d represent the source and driving 3D implicit keypoints, respectively, while x_c,s denotes the canonical keypoints of the source image.

Thanks for your reply. I‘m sorry, I made a spelling mistake. My questiones about formula (3), not formula (2). @zzzweakman

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants