-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to reproduce published results #13
Comments
I've been trying to reproduce the results for a long time, however, I still cannot get it, either. I hope the authors could provide the exact model architecture and hyper-parameters. And also, I found that the model in the MNIST(2000)->USPS expriment is not as same as the CCSA keras implementation. |
@KevinDocel, you are right about the model not being the same as the one in CCSA. To be fair, I have implemented it in d-SNE's codebase. Please see commit f98083. |
@sheikhomar, the differences not only lie in the feature dimension as you did, but also the first 2 conv layers, including the kernel size and channels. BTW, the input size to CCSA keras model is 16x16, while it seems that 32x32 is used in the authors' implementation. But, simply changing the input size from 32x32 to 16x16 in CCSA keras model can not reproduce the results, since this will lead to heavily overfitting for CCSA keras model as I've tried. |
Hi, |
Hi @ShownX. Thank you for taking the time to develop a new starter codebase. It contains a functional MNIST -> MNIST-M test (using the
I modified the MAX_EPOCH to 5 for curiosity's sake and ran a test. I have attached the output in a file -- if you have a moment, could you verify that these values are as expected? Additionally, looking at the table in the arXiv publication for MT-MM, for d-SNE with 10 target images/class, it appears I should get ~88% accuracy. However, in the attached log, I noticed that the accuracy seems to stall around 11% at the beginning, then abruptly starts to increase for a number of iterations, then overfits. (See: Test accuracy: 81%, Target/Source accuracy: 100%/99%) This loss/accuracy evolution seems like it might not be ideal. I am wondering about possible concerns about initialization for the network (could be the cause of stalling at 11% at beginning.) I am also wondering about possible concerns about the chosen hyperparameters (could be the cause of sharp increase then overfitting.) Do you have any information about the configuration used to produce the 88% reported accuracy? Also, for the Thank you kindly for your time and effort. :) |
@sheikhomar How did you get MNIST-M and other digits datasets? In your loader I only see code for MNIST and USPS |
The original source of the MNIST-M dataset can be found on Yaroslav Ganin's page under the "Unsupervised Domain Adaptation by Backpropagation" subheading. Be sure to use the "unpacked version of MNIST-M" link, as the packed version of the dataset contains LMDB files that need to be read using the lmdb library, which d-SNE does not use. The SVHN dataset can be found here, as well. |
Hi,
Thank you for a great paper and for publishing the code for d-SNE.
I am trying to reproduce the MNIST-USPS results in Table 1 in your paper but so far I have been unable to do so.
To reproduce the results I have done following this:
./scripts/digits-mt-us-su-s.sh
./scripts/digits-mt-us-su-s.sh
again. Please see the PR for the changes I made.In my analysis, I get the following Eval-Acc-Tgt accuracies when performing domain adaptation from MNIST to USPS:
What did I do wrong? Can you please provide some guidance?
The text was updated successfully, but these errors were encountered: