-
Notifications
You must be signed in to change notification settings - Fork 493
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: enable self supervised pretraining #220
Conversation
Hi again! Big fan of:
I'd suggest:
For me the ideal workflow here is that saved models will have trained embedding + encoder; may have a decoder; and may have one or more output/decision modules - depending on whether they've been pre-trained & what supervised task(s) they've been trained on. For example we could use a I would like to see the training procedure (e.g. *Model) use None/NaN as the mask value so it looks the same as actual missing data, and have missing-value featurization done (configurably) in
In a small preliminary test (on Forest Cover Type again 😂 - just masking some values at random and doing supervised classification task), I actually found that this extra column featurization delivered best results despite significantly increasing the Thanks again & LMK your thoughts! |
@athewsey interesting. Lots of information at the same time, but definitely some nice feature proposals. Maybe you could open one separate issue for each feature (nans, loading only decoder etc...) to ease the discussion. |
2f725ba
to
4f4ffff
Compare
chore: linting, fix variables and format wip: pretraining notebook WIP: pretraining almost working feat: add self supervision
4f4ffff
to
e88dc38
Compare
chore: fix lint chore: update README feat: add explain to unsupervised training feat: update network parameters When the network is already defined we still need to update some parameters fed through the fit function such as the virtual batch size and, in the case of unsupervised pretraining, the pretraining_ratio.
09ce9d1
to
6a2826c
Compare
What kind of change does this PR introduce?
Implements self supervised pretraining
Does this PR introduce a breaking change?
I would say so
What needs to be documented once your changes are merged?
Some documentation
Closing issues
Closes #187 #232