Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What dataset is used to train the pretrained model? #17

Closed
xieenze opened this issue Sep 2, 2019 · 27 comments
Closed

What dataset is used to train the pretrained model? #17

xieenze opened this issue Sep 2, 2019 · 27 comments

Comments

@xieenze
Copy link

xieenze commented Sep 2, 2019

Hello author:
I try to train ASTER on Synth800K and Synth90K, same setting as the TPAMI Papaer. However, I can not reimplement the accuracy in the paper and your pretrained model.
Can you tell me what dataset are you used to train the pretrained model you provided? Thanks very much!

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 3, 2019

Both the pytorch port and the tensorflow port are trained with Synth90k and SynthText, just as the paper described. Have you tried the tensorflow version?

@xieenze
Copy link
Author

xieenze commented Sep 3, 2019

No, I didn't run TensorFlow version. But this PyTorch I use Synth and Syn90K, I failed to run the accuracy you reported in readme.md.
My accuracy is average 1.5-2 lower than yours in all datasets.
Can you tell me some tricks or details maybe I ignored to train such high accuracy ASTER?
Thanks very much!

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 3, 2019

Actually I do not use any tricks to achieve such performance. All results reported in the README.md are trained and tested with the given scripts.

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 3, 2019

Hi @xieenze,

With the released model, can you obtain the reported results in the README.md.

@xieenze
Copy link
Author

xieenze commented Sep 3, 2019

Yes, using your released model, I can achieve the same results in readme
Maybe I need to check sth. Because I use 4 gpus and modify the batch size and lr.
Thanks very much!

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 6, 2019

Hi enze,

Have you reproduced the reported results? If yes, please let me know, I am curious about the effects from different hyper-parameters.

@WenjiaWang0312
Copy link

WenjiaWang0312 commented Sep 6, 2019

Hi, I used the workers of 100, and batch_size of 1024. I also modified the learning-rate by hand when the accuracy stuck at a point. I had not achieved the accuracies on those 7 test datasets as your report. Approximately 1.0% to 2.0% lower than your results.

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 7, 2019

emmmm... why not try the given settings? I can not guarantee the same performance with different settings. @JasonBoy1

@WenjiaWang0312
Copy link

emmmm... why not try the given settings? I can not guarantee the same performance with different settings. @JasonBoy1

OK, I should try recently

@WenjiaWang0312
Copy link

emmmm... why not try the given settings? I can not guarantee the same performance with different settings. @JasonBoy1

thank you very much

@WenjiaWang0312
Copy link

Hi, dear author. I found that the default batch_size in the script/stn_att_rec.sh is different from that in config.py. Which did you used in your experiment?
And how many GPUs did you used?

@ayumiymk
Copy link
Owner

ayumiymk commented Sep 8, 2019

The arguments in config.py are just default values for definition. Actually, you should just run bash script/stn_att_rec.sh. All the settings are given in this script, including the number of GPUs (here 2 GPUs are used).

@minglii1998
Copy link

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

@ayumiymk
Copy link
Owner

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

Maybe you can try our tensorflow implement. If you can not obtain the same results using both the tf and pytorch version. I suspect that there is something wrong with your training data.

@minglii1998
Copy link

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

Maybe you can try our tensorflow implement. If you can not obtain the same results using both the tf and pytorch version. I suspect that there is something wrong with your training data.

Thanks for reply~
Still I can't achieve the performance. I wonder if there's any chance you could provide me with the datasets in lmdb format, I would be really grateful~
Sorry to bother!

@ayumiymk
Copy link
Owner

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

Maybe you can try our tensorflow implement. If you can not obtain the same results using both the tf and pytorch version. I suspect that there is something wrong with your training data.

Thanks for reply~
Still I can't achieve the performance. I wonder if there's any chance you could provide me with the datasets in lmdb format, I would be really grateful~
Sorry to bother!

How about the performance of the Tensorflow implement?

What's more, I am sorry. The size of the training data is about 50GB. I don't have such a big cloud drive. You should carefully check your own data.

@minglii1998
Copy link

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

Maybe you can try our tensorflow implement. If you can not obtain the same results using both the tf and pytorch version. I suspect that there is something wrong with your training data.

Thanks for reply~
Still I can't achieve the performance. I wonder if there's any chance you could provide me with the datasets in lmdb format, I would be really grateful~
Sorry to bother!

How about the performance of the Tensorflow implement?

What's more, I am sorry. The size of the training data is about 50GB. I don't have such a big cloud drive. You should carefully check your own data.

Thanks anyway, I will try the tf implement~

@ayumiymk
Copy link
Owner

@minglii1998 If you make any progress, please let me know. Thanks!

@minglii1998
Copy link

Hello there, finally I achieved the the performance as @ayumiymk mentioned, thus it's for sure that there's nothing wrong with this code.
The performance is almost the same, maybe 0.1-0.3 lower or higher like that.
The reason I can't get the same acc mostly because the datasets I used. The way I pre-process the ST dataset is not right. Handling the exception is of great importance while getting the lmdb format data.

@WenjiaWang0312
Copy link

WenjiaWang0312 commented Oct 12, 2019

Hello there, finally I achieved the the performance as @ayumiymk mentioned, thus it's for sure that there's nothing wrong with this code.
The performance is almost the same, maybe 0.1-0.3 lower or higher like that.
The reason I can't get the same acc mostly because the datasets I used. The way I pre-process the ST dataset is not right. Handling the exception is of great importance while getting the lmdb format data.

Dear MingLii:
Could you tell me what's the right way to 'handling the exception' of the ST dataset? My implemented accuracy is about 2%-3% lower than the author's report.

@Pay20Y
Copy link

Pay20Y commented Oct 17, 2019

I wonder how many images are cropped from ST800K? Thanks! @minglii1998

@ofexe
Copy link

ofexe commented Oct 17, 2019

@ayumiymk Could you share the scripts that you used to create lmdb from Synth90k and SynthText datasets? That would be helpful, thank you.

@WenjiaWang0312
Copy link

Dear author:
Could you share what's the right way to 'handling the exception' of the ST dataset? Like the @minglii1998 said.

@minglii1998
Copy link

@JasonBoy1 @Pay20Y @ofexe
Apologies for my not reply, I was so busy that even if I received the email from git, I still couldn't show up. The author shared the datasets he used with me, and I will share them here with his permission.
You can download from from here,code:wi05
Hope all of all of you can get a good result!
@ayumiymk Hi bro, maybe it's about time to close this issue, thanks for your sharing!

@Pay20Y
Copy link

Pay20Y commented Oct 19, 2019

@minglii1998 thanks a lot!

@ofexe
Copy link

ofexe commented Oct 22, 2019

@minglii1998 I've already made my own datasets, but still thanks.

@YooWang
Copy link

YooWang commented May 26, 2020

Hello there, I'd like to know if any one of you could achieve the same performance ~
I ran bash script/stn_att_rec.sh with two GPUs, but it seemed that my accuracy is about 1-2 lower than the ASTER (L2R) and much lower than ASTER.Pytorch, which should I expect?

Maybe you can try our tensorflow implement. If you can not obtain the same results using both the tf and pytorch version. I suspect that there is something wrong with your training data.

Thanks for reply~
Still I can't achieve the performance. I wonder if there's any chance you could provide me with the datasets in lmdb format, I would be really grateful~
Sorry to bother!

How about the performance of the Tensorflow implement?
What's more, I am sorry. The size of the training data is about 50GB. I don't have such a big cloud drive. You should carefully check your own data.

Thanks anyway, I will try the tf implement~

hello, can you teach me how training from begin.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants