You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Some models make use of some sort of scheduling or annealing internally (e.g. KL warmup or temperature annealing) based on the current step index - what's the correct way to implement this within pythae?
The text was updated successfully, but these errors were encountered:
Thank you for opening this issue. Yes indeed some models may need several warmup steps during training. To allow such a behavior, the current training epoch is automatically passed to the forward method of the model in the trainers as shown below
The current epoch can be retrieved in the forward method as follows
epoch=kwargs.pop("epoch")
and then passed to the loss_function. Nonetheless, this warmup setting is only supported for some models for now and you may require to make the changes in the model implementation yourself. If you do so, you will need to first clone the repo and install it in editable mode
Some models make use of some sort of scheduling or annealing internally (e.g. KL warmup or temperature annealing) based on the current step index - what's the correct way to implement this within
pythae
?The text was updated successfully, but these errors were encountered: