Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

tuning the downstream MLPF model [pytorch] #166

Merged
merged 13 commits into from
Feb 6, 2023
Merged

tuning the downstream MLPF model [pytorch] #166

merged 13 commits into from
Feb 6, 2023

Conversation

jpata
Copy link
Owner

@jpata jpata commented Feb 1, 2023

  • switch MLPF optimizer to AdamW [pytorch]
  • use focal loss [pytorch]
  • separate decoding layers for different regression components [pytorch]
  • added main19.cc for generating CLIC events with PU
  • switch CLIC jet clustering to ee_genkt [TF] [pytorch]

@jpata jpata changed the title Feb 2023: tuning the models Feb 2023: tuning the downstream MLPF model [pyg] Feb 4, 2023
@jpata
Copy link
Owner Author

jpata commented Feb 4, 2023

Stopped due to patience=20.

jet_res

Screenshot 2023-02-04 at 19 46 25

Resubmitted with higher dropout, patience=50.

python3 ssl_pipeline.py --data_split_mode domain_adaptation \
  --prefix_VICReg pytorch_${SLURM_JOB_ID} --prefix_mlpf MLPF_test \
  --train_mlpf --native --n_epochs_VICReg 0 --batch_size_mlpf 100 \
  --n_epochs_mlpf 500 --patience 50 --width_mlpf 512 --embedding_dim 512 --lr 0.00005 --nearest 32 --evaluate

@jpata
Copy link
Owner Author

jpata commented Feb 6, 2023

Choosing the best (gray line) out of 3 trainings with the same hyperparameters, I'm able to get the pytorch MLPF model to reproduce native PF (and the TF model) in terms of jet response.

Screenshot from 2023-02-06 10-37-28

jet_res

This was run with the following options:

  python3 ssl_pipeline.py --data_split_mode domain_adaptation \
  --prefix_VICReg pytorch_${SLURM_JOB_ID} --prefix_mlpf MLPF_test \
  --train_mlpf --native --n_epochs_VICReg 0 --batch_size_mlpf 100 \
  --n_epochs_mlpf 500 --patience 50 --width_mlpf 256 --embedding_dim 256 --lr 0.00005 --num_convs_mlpf 3 --nearest 32 --evaluate

@jpata jpata changed the title Feb 2023: tuning the downstream MLPF model [pyg] Feb 2023: tuning the downstream MLPF model [pytorch] Feb 6, 2023
@jpata jpata changed the title Feb 2023: tuning the downstream MLPF model [pytorch] tuning the downstream MLPF model [pytorch] Feb 6, 2023
@jpata jpata merged commit 4846e25 into main Feb 6, 2023
@jpata jpata deleted the feb23_dev branch August 28, 2023 15:33
jpata added a commit that referenced this pull request Sep 15, 2023
* fix simulation seed setting

* tuning the models

* add focal loss, ee jet algo

* run sim with PU

* tuning

* increase dropout
jpata added a commit that referenced this pull request Sep 15, 2023
* fix simulation seed setting

* tuning the models

* add focal loss, ee jet algo

* run sim with PU

* tuning

* increase dropout

Former-commit-id: 44c4d7d
jpata added a commit that referenced this pull request Sep 15, 2023
* fix simulation seed setting

* tuning the models

* add focal loss, ee jet algo

* run sim with PU

* tuning

* increase dropout

Former-commit-id: 44c4d7d
jpata added a commit that referenced this pull request Sep 25, 2023
* fix simulation seed setting

* tuning the models

* add focal loss, ee jet algo

* run sim with PU

* tuning

* increase dropout
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant