Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add minimal pretrained checkpoints for tests and fix PAF grouping interpolation #73

Merged
merged 14 commits into from
Sep 12, 2024
Merged
197 changes: 107 additions & 90 deletions docs/config.md

Large diffs are not rendered by default.

169 changes: 37 additions & 132 deletions docs/config_bottomup.yaml
Original file line number Diff line number Diff line change
@@ -1,118 +1,36 @@
data_config:
provider: LabelsReader
pipeline: BottomUp
train:
labels_path: minimal_instance.pkg.slp
train_labels_path: minimal_instance.pkg.slp
val_labels_path: minimal_instance.pkg.slp
preprocessing:
max_width: null
max_height: null
scale: 1.0
is_rgb: false
preprocessing:
crop_hw:
- 160
- 160
augmentation_config:
random_crop:
random_crop_p: 0
random_crop_hw:
- 160
- 160
use_augmentations: true
augmentations:
intensity:
uniform_noise:
- 0.0
- 0.04
uniform_noise_p: 0
gaussian_noise_mean: 0.02
gaussian_noise_std: 0.004
gaussian_noise_p: 0
contrast:
- 0.5
- 2.0
contrast_p: 0
brightness: 0.0
brightness_p: 0
geometric:
rotation: 180.0
scale: 0
translate:
- 0
- 0
affine_p: 0.5
erase_scale:
- 0.0001
- 0.01
erase_ratio:
- 1
- 1
erase_p: 0
mixup_lambda: null
mixup_p: 0
val:
labels_path: minimal_instance.pkg.slp
max_width: null
max_height: null
is_rgb: false
scale: 1.0
preprocessing:
crop_hw:
- 160
- 160
augmentation_config:
random_crop:
random_crop_p: 0
random_crop_hw:
- 160
- 160
use_augmentations: false
augmentations:
intensity:
uniform_noise:
- 0.0
- 0.04
uniform_noise_p: 0
gaussian_noise_mean: 0.02
gaussian_noise_std: 0.004
gaussian_noise_p: 0
contrast:
- 0.5
- 2.0
contrast_p: 0
brightness: 0.0
brightness_p: 0
geometric:
rotation: 180.0
scale: 0
translate:
- 0
- 0
affine_p: 0.5
erase_scale:
- 0.0001
- 0.01
erase_ratio:
- 1
- 1
erase_p: 0
mixup_lambda: null
mixup_p: 0
use_augmentations_train: true
augmentation_config:
geometric:
rotation: 180.0
scale: null
translate_width: 0
translate_height: 0
affine_p: 0.5

model_config:
init_weights: xavier
pre_trained_weights: null
backbone_type: unet
backbone_config:
backbone_type: unet
backbone_config:
in_channels: 1
kernel_size: 3
filters: 16
filters_rate: 2
max_stride: 16
convs_per_block: 2
stacks: 1
stem_stride: null
middle_block: true
up_interpolate: true
in_channels: 1
kernel_size: 3
filters: 16
filters_rate: 2
max_stride: 16
convs_per_block: 2
stacks: 1
stem_stride: null
middle_block: true
up_interpolate: true

# pre_trained_weights: ConvNeXt_Tiny_Weights
# backbone_config:
Expand Down Expand Up @@ -144,20 +62,20 @@ model_config:
# stem_patch_stride: 2

head_configs:
confmaps:
head_type: MultiInstanceConfmapsHead
head_config:
part_names: None
sigma: 1.5
output_stride: 2
loss_weight: 1.0
pafs:
head_type: PartAffinityFieldsHead
head_config:
edges: None
sigma: 50
output_stride: 4
loss_weight: 1.0
single_instance:
gitttt-1234 marked this conversation as resolved.
Show resolved Hide resolved
centered_instance:
centroid:
bottomup:
confmaps:
part_names: None
gitttt-1234 marked this conversation as resolved.
Show resolved Hide resolved
sigma: 1.5
output_stride: 2
loss_weight: 1.0
pafs:
edges: None
gitttt-1234 marked this conversation as resolved.
Show resolved Hide resolved
sigma: 50
output_stride: 4
loss_weight: 1.0
trainer_config:
train_data_loader:
batch_size: 4
Expand All @@ -169,7 +87,6 @@ trainer_config:
model_ckpt:
save_top_k: 1
save_last: true
device: cpu
trainer_devices: 1
trainer_accelerator: cpu
enable_progress_bar: false
Expand All @@ -179,18 +96,6 @@ trainer_config:
use_wandb: false
save_ckpt: true
save_ckpt_path: min_inst_bottomup1
wandb:
entity: team-ucsd
project: test_centroid_centered
name: fly_unet_centered
wandb_mode: ''
api_key: ''
log_params:
- trainer_config.optimizer_name
- trainer_config.optimizer.amsgrad
- trainer_config.optimizer.lr
- model_config.backbone_config.backbone_type
- model_config.init_weights
optimizer_name: Adam
optimizer:
lr: 0.0001
Expand Down
Loading
Loading