Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the training setting #4

Open
zht022 opened this issue Jan 29, 2020 · 1 comment
Open

Question about the training setting #4

zht022 opened this issue Jan 29, 2020 · 1 comment

Comments

@zht022
Copy link

zht022 commented Jan 29, 2020

Hi, @LiangSiyuan21 , thank you for sharing your training code. I've tried to train the GAN from scratch with this code but got confused.

I've noticed that the implementation of your generator is not the same as you mentioned in the paper (encoder-decoder), while the discriminator is the same as [pix2pix] but not the ResNet-32 mentioned in the paper.

Meanwhile, other parameters setting is almost the same as the paper: (1) objective function's weight \alpha=0.05, \beta=1, \epsilon=1e-4&2e-4;
(2) use Adam solver with a learning rate of 0.0002 for D (but 0.0001 for G), and momentum parameters \beta1 = 0.5, \beta2 = 0.999;
(3) divide the objective by 2 while optimizing D;
(4) weights were initialized from a Gaussian distribution with mean 0 and standard deviation 0.02;
(5) for each input image, min_size & max_size = 300.
Note that (3), (4), (5) are referred from [pix2pix].

However, after 3 training epoch, I've discovered that:
(1) The attacking result is not good enough.
(2) The adv_images are strange. Distance between real & fake are not very closed, the objects in the adv_images are often with green shadows around them. Some examples are given here:

image

image
image
image

My docker environment: python 3.7, pytorch 1.2.0, cuda 10.0, cudnn 7, cupy-cuda100.

If there exists any mistake in the above settings, please figure it out. What's more, I'm wondering how to check the correctness of the training process, if you can provide a training log, I would greatly appreciate!

@DuanXiaoyue-LittleMoon
Copy link

Hello, I'm wondering whether you've meet with the problem that the misclassification loss in these 4 losses is always 0 during the training process and how you solved it? Thanks very much.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants