You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.
Hi, I have been reading your paper and code,
and I am confused about how the best model of the entire training process is selected.
this is how I understood the training code
model training (both burn-in and mutual learning stage) is performed on train data
model weight is saved every 5000 steps, by hooks.PeriodicCheckpointer
After the last training step is finished (MAX_ITER reached), resulting weight is used for evaluation
Please correct me if i am wrong.
and my questions are:
a. Should I take the model weight after the last training step as the final model weight for future inference?
b. It seems validation loss/metric is not calculated in the code, but in the paper there is a plot of validation mAP (Figure 4 )
Are the metrics reported on the paper calculated with post last training step weights or weight selected based on validation set?
c. Is there a model selection based on validation loss/metric function that i missed in this repo?
Thank you for the great paper and code
I found the contents really interesting.
Thanks in advance!
The text was updated successfully, but these errors were encountered:
a. I usually test the model weight of the middle training scenario which has better performance.
b. I calculate them in tensorboard. I downloaded the csv file and draw the curve using python matplotlib.
c. I select the best model from the tensorboard curve and go back to search for the saved checkpoint.
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi, I have been reading your paper and code,
and I am confused about how the best model of the entire training process is selected.
this is how I understood the training code
Please correct me if i am wrong.
and my questions are:
a. Should I take the model weight after the last training step as the final model weight for future inference?
b. It seems validation loss/metric is not calculated in the code, but in the paper there is a plot of validation mAP (Figure 4 )
Are the metrics reported on the paper calculated with post last training step weights or weight selected based on validation set?
c. Is there a model selection based on validation loss/metric function that i missed in this repo?
Thank you for the great paper and code
I found the contents really interesting.
Thanks in advance!
The text was updated successfully, but these errors were encountered: