Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The sigloss does not go down for custom dataset #54

Closed
prateekstark opened this issue Sep 17, 2022 · 2 comments
Closed

The sigloss does not go down for custom dataset #54

prateekstark opened this issue Sep 17, 2022 · 2 comments

Comments

@prateekstark
Copy link

prateekstark commented Sep 17, 2022

Hi there, I tried using the DepthFormer model for my custom dataset but surprisingly the sigloss in the validation set does not reduce. I referred to #20 but I noticed this is not a random phenomenon, it happens everytime. I tried restarting many times. Could you point out some possible solution? I also tried the warming up of SIGLOSS, doesn't help. For my case, I chose:
depth_scaling: 1000 for 16 bit depth images.
Min, Max depth: 1e-3 and 10.
Image Size: (512, 384)
Configs I tried for: DepthFormer, BinsFormer

Rest all Parameters are same as respective NYU configs

@zhyever
Copy link
Owner

zhyever commented Sep 17, 2022

That sucks.

Please first ensure that the value of your depth maps lies in 0-10 after dividing it by the depth_scale=1000.

Then, try to set scale_up=True presented in issue#20 to modify the depth head into a sigmoid version.

Actually, in the experiments of my new paper LiteDepth, I still observe this convergence issue. It is so hard to find the reason and I also observe this issue in other repos like TransDepth and Adabins. Hence, it may be related to general experimental settings like crop size, color aug, sigloss itself, and so on. I also look for help from the community to fix this bug.

Anyway, a more direct way to solve this issue: loading our pre-trained model. You may see some warning about misalignment of param_dict, but that's ok and will not break down the training since I set the strict=False of load_checkpoint. Since the model is pre-trained on the depth task, it can avoid breaking down in the early stage of training. Hope it is useful for you.

Sometimes, I also think it is not a random phenomenon. But after repeating re-starting several times and setting different random seeds, it works... Gook luck! :D

@zhyever
Copy link
Owner

zhyever commented Nov 29, 2022

Close it for now.

@zhyever zhyever closed this as completed Nov 29, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants