You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In that case, the encoder is random... and this is quite unexpected.
Potential solutions:
Add a mechanism in experimaestro that allows linking the sub-models to the main one - when initialized, the encoder will require the model to be loaded from the checkpoint. Caveats: as with any implicit process, this makes xpmir more complex and might have other side effects.
Add a mechanism to have several tasks to be run (and in that case, the output will be just the path), instead of just one, i.e.
Important caveat: it breaks the current HuggingFace uploading/downloading mechanism and we need to include load_model each time we want to use the pre-trained model (which can begin to be cumbersome when various models are combined since we need to track down this). For HF, the solution would be to have this when loading
When a model is trained, one might want to access one of its submodules in another task, e.g.
In that case, the encoder is random... and this is quite unexpected.
Potential solutions:
Add a mechanism in experimaestro that allows linking the sub-models to the main one - when initialized, the
encoder
will require themodel
to be loaded from the checkpoint. Caveats: as with any implicit process, this makesxpmir
more complex and might have other side effects.Add a mechanism to have several tasks to be run (and in that case, the output will be just the path), instead of just one, i.e.
Important caveat: it breaks the current HuggingFace uploading/downloading mechanism and we need to include
load_model
each time we want to use the pre-trained model (which can begin to be cumbersome when various models are combined since we need to track down this). For HF, the solution would be to have this when loadingThird solution, do something a bit in between
where
model_loader
is a specialexperimaestro
object (ConfigFactory
) that does whatever is needed and returns the "encoder" submodel.The text was updated successfully, but these errors were encountered: