Replies: 2 comments
-
Hi @hz1z, thanks for your interest here.
In MONAI, all transforms assume input data shape as [channel x spatial_dims].
Using random clipping to a smaller size (96x96x96) during the training phase and a larger size (160x160x160) during the inference phase is a common strategy that aims to balance the generalization ability of the model and the accuracy in real-world applications. With this approach, the model is able to learn richer features during training and make full use of them during inference to improve performance. convert to discussion more like a usage issue, feel free to create another if meet any other issue. Thanks! |
Beta Was this translation helpful? Give feedback.
-
thank you ! @KumoLiu As for the second issue, I seem to have grasped it; it might be related to the local characteristics of CNNs. It is possible to train on smaller patches and infer on larger patches without affecting the network accuracy. |
Beta Was this translation helpful? Give feedback.
-
in you tutorial
spleen_segmentation_3d
question1
Isn’t this redundant? This code decollates the data from the batch dimension and then wraps it into a list, effectively doing nothing, right? Why not directly use post_pred(val_outputs) and post_label(val_labels)? Does it mean that the AsDiscrete transform has to remove the data from the batch dimension to work?
question2
why roi_size = (160, 160, 160)?
for training,you randomly crop the data used to a size of (96, 96, 96)? Why are you using (160, 160, 160) for inference instead?
Beta Was this translation helpful? Give feedback.
All reactions