Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

investigate why the ap not good in voc dataset #63

Open
xytpai opened this issue Jun 23, 2019 · 4 comments
Open

investigate why the ap not good in voc dataset #63

xytpai opened this issue Jun 23, 2019 · 4 comments

Comments

@xytpai
Copy link

xytpai commented Jun 23, 2019

I'v heard many achieved 60+ mAP only use voc0712-trainval for training, and I got similar result in my experiment. The mAP during training are shown in the figure. I'v achieved 68 mAP with data augment (random resize crop) and use 641*641 as input.
res
I think the main reason is that the labeling of VOC dataset is not accurate enough.
Since Retinanet takes all the pixels that are not near the target box as background, the labeling accuracy is important. But in VOC we found some objects are not labeled.
res2

@sunshine-zkf
Copy link

I think the VOC dataset is not a problem and should be a problem with code. I tested the trained model, and all the pixels that are not near the target box.

@xytpai
Copy link
Author

xytpai commented Jun 27, 2019

I think the VOC dataset is not a problem and should be a problem with code. I tested the trained model, and all the pixels that are not near the target box.

Thanks for reply, I examined the code carefully and found several differences from the official code and finally reached 76 map. There is nothing to do with voc. The most critical issue is that the author set the threshold to 0.5, which is actually 0.05.

@07hyx06
Copy link

07hyx06 commented Nov 1, 2019

hi! What's your final loss in training dataset? Should i set CLS_THRESH=0.5 and LOC-THRESH=0.05?
i get about 0.02 train loss but lots of bounding box appear in the test image :(

@RenzhiDaDa
Copy link

RenzhiDaDa commented Dec 11, 2019

I agree with @xytpai. When I check the code ,i find the encode.py has the difference between the paper and the code.the paper sets the CLS_THRESH =0.05,but this code uses 0.5.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants