-
Notifications
You must be signed in to change notification settings - Fork 79
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Weird results when training with provided script on RobotCar loop scene. #43
Comments
Hi @KongYuJL is it an issue with
|
|
Thanks for your suggestions! And when I trained on the 7Scenes dataset, there is no problem and the results are very similar to your reported performance. |
@KongYuJL you can also see if the training loss values match with the pretrained model we provided. |
Hello, pardon my presumptuousness. I have encountered the same problem as yours, how did you solve it in the end? |
Well, it's been a while, I don't know if my memory is accurate. |
Thanks for your answer. But if it is a problem of inconsistency in the environment, why is it successful on the 7Scenes dataset, and why does it fail on RobotCar? One difference between the two datasets is that RobotCar uses the SDK, could this be the reason? |
@cccccv-cm The data range of RobotCar is different from the 7SCenes, maybe that causes some problems in the training process. Especially there are some differences when using python2 and python3? I'm not sure about this. It's really wired work in 7Scenes but failed in RobotCar, I tried to find the reason. But cannot find the related codes. I suspect the CUDA version may also be one of the potential reasons. |
Hi there, @samarth-robo.
Thanks for your solid work and your well-structured code, I could produce even better results than the numbers in the original paper by loading pre-trained model weights.
However, when I retrained on the RobotCar dataset and loop scene by using the provided script and config file (from latest version):
python train.py --dataset RobotCar --scene loop --config_file configs/mapnet.ini --model mapnet --device 1 --learn_beta --learn_gamma
The results become weird and errors are much larger than I expected.
It's worth noting that, I executed the script on an 8 * NVIDIA RTX 2080ti node:
When I used pytorch-0.4.1, which is specified in your environment.yaml, there was an error detected by cuda: "THCudaCheck FAIL file=/pytorch/aten/src/THC/THCGeneral.cpp line=663 error=11 : invalid argument"
Thus I ran the script both in pytorch 0.4.1 and 1.0.1 environment. However, both errors are very large.
Besides, I also noticed that the preprocessed images have some over-exposure cases (some are almost all white and barely has information), is it normal?
like 1403774724292807.png, ... 1403774724917727.png at the beginning of 2014-06-26-09-24-58 and other sequences.
The text was updated successfully, but these errors were encountered: