Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use your own pre-trained model for downstream segmentation #204

Open
yyy1998-i opened this issue Aug 28, 2024 · 6 comments
Open

Use your own pre-trained model for downstream segmentation #204

yyy1998-i opened this issue Aug 28, 2024 · 6 comments

Comments

@yyy1998-i
Copy link

I would like to ask you a question, I pre-train on my medical image data set (6k) according to the official pre-training code, and then the downstream task is segmentation, but the fine-tuning results with my own pre-training model is always inferior to the author's pre-training model with imagenet, which is why, is my pre-training data is still too small?

@Jivitesh2001
Copy link

Hey,
Your dataset seems too small compared to Imagenet-1K. A suggestion - you should initialize your model using the pre-trained model weights of Imagenet and then continue pre-training on your medical image dataset, which you can use for your downstream task. Please let me know how this goes for you.

@yyy1998-i
Copy link
Author

Hey, Your dataset seems too small compared to Imagenet-1K. A suggestion - you should initialize your model using the pre-trained model weights of Imagenet and then continue pre-training on your medical image dataset, which you can use for your downstream task. Please let me know how this goes for you.

Ok, thank you very much for your reply. Next, I will try according to what you said. Thanks again

@Jivitesh2001
Copy link

Hello,
Have you tried this approach? and Can you tell me what were your results?

@yyy1998-i
Copy link
Author

Hello, Have you tried this approach? and Can you tell me what were your results?

Sorry for the late reply. This is my latest training situation, and it is still not as good as the weight of imagenet.
7c58737620d28084707bb3ea6c8c735f

@Jivitesh2001
Copy link

Hey!
Have you tried different augmentation techniques to oversample your medical dataset? I think the idea should be a) Load the ImageNet Pre-Trained Checkpoint, b) Then use the pre-trained checkpoint to fine-tune the model (which seems to be a classification model) on your dataset for your downstream task. Have you tried this approach?

@yyy1998-i
Copy link
Author

Hey! Have you tried different augmentation techniques to oversample your medical dataset? I think the idea should be a) Load the ImageNet Pre-Trained Checkpoint, b) Then use the pre-trained checkpoint to fine-tune the model (which seems to be a classification model) on your dataset for your downstream task. Have you tried this approach?

Next, I will try to use data enhancement (the previous training was in accordance with the official default Settings). In the pre-training process, I found that without using normalize,MAE's reconstruction effect is better, and the loss is also very low (0.0822). When fine-tuning, I used mean and std which I used to fine-tune the data set. I found that this also affected the effectiveness of the downstream classification task.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants