-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Recommended] Questions on Re-implementation #3
Comments
Thank you for your affection for HifiFace, this project is still undergoing the approval process of open source. Thank you for your patience. Additionally we sincerely hope you can raise your question again in English and these information can be shared by more of this community.
The size of
Feedbacks for question 3 are listed below. |
For questions in part 3,
|
For questions in part 3,
|
Thanks a lot for your response. It is very helpful to understand the paper better. Some questions about dataset:
Some questions about training:
Some questions about implementation:
|
About Section 5
About Section 6
|
About Section 4
|
What is the effect to the generator and the synthesis result whether to use InstanceNorm in Dicriminator? |
Sorry for a late reply. It is quite weird that you encounter this issue. In fact, in our implementation, we do not use instance norm in the discriminator. In the generator, besides AdaIn, we use InstanceNorm in the encode and bottleneck part. I wonder if your BP of D loss is correct. Remember to detach the generator part when BP the D loss. |
Yeah, I fixed some tricks and the discriminator without IN worked now, but there seems no difference with the discriminator using IN. I really wonder the effect of IN on the discriminator, and single output v.s. PatchGAN. I'd be very happy to learn your opinion! |
HifiFace is mainly researched on image based swapping instead of video. So it is normal that the model performs not that perfect in video, as we did not tune this. If you need any help on video synthesis, feel free to drop me an email and maybe we can provide an official generation for you to compare with your result. |
Hi, I've send an email to hififace.youtu@gmail.com, wish for your reply! Thanks! |
Hi, according to what you mentioned above, |
Your job is great and really nice.
And in my experimentation, PatchGAN seems to perform better. |
Exactly |
Thank you very much for your advice, we will definitely try more SOTA backbones for better results! |
Thx! I still have some questions:
|
|
Hi, author, thanks your great job! I have some questions about implementation details:
|
I try to re-implement the paper,but meet some troubles.Ask for help. |
How did you apply color re-normalization? Is there any reference article or code? |
您好,我在复现您的这篇论文,在复现过程中遇到了以下问题:
从数据集Celebrity-Asian和VggFace2中裁剪人脸,存在很多模糊人脸,请问数据集中的模糊数据您是如何处理的,您最终用于训练模型的数据量大概是多少,这部分在论文中没有详细提及;
在Feature-Level最后一段中:
在Experiments中:
再结合您在issue2中的回答,您看我下面的理解是否正确:
2.1 Ours-256 和 Ours-512 模型输入的 It 均为256尺寸图像;
2.2 在 Ours-256 模型中,在获得 zfuse 后,先过两层 AdaIN Res-Block,再过两层有upsample的Res-Block的Fup;
2.3 Ours-512 与 Ours-256 相比,区别仅在于,在Ours-512模型的 Fup模块中多一层有upsample的Res-Block;
3.1 获得 Mlow 、Mr、Ilow、Ir过程中,输出层结构是怎样的;
3.2 使用3D人脸模型获得人脸关键点后,其坐标范围为[0, 224],是否将其转换为[0, 1]范围;
3.3 在计算 Lcyc 时,cyc过程输出G(Ir , It),Ir是否有detach;训练过程中发现Lcyc 特别低,相比其他loss差一到两个数量级;
3.4 训练过程中是怎样做数据采样的;
3.5 在辨别器中使用的也是Res-Block,Res-Block中使用的是InstanceNorm2D,这里请您确认在辨别器中使用的是InstanceNorm2D;
3.6 能否列举SFF的详细结构;
3.7 HRNet的人脸分割效果比较差,是否做了其他优化;
3.8 是否是分阶段训练的,还是在一开始辨别器就参与了训练;
3.9 脸型差异较大时,生成结果存在双下巴现象,这种现象是否是靠辨别器抑制掉的。
问题较多,期待回复,谢谢。
The text was updated successfully, but these errors were encountered: