You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Ah yes, that script only support causal lms right now. A note on your solution, I'm not certain, but AutoModel here may give you a T5Model rather than a T5ForConditionalGeneration as you may want. Probably worth double checking that.
Just double checked and the model was indeed marked as a T5Model and not as a T5ForConditionalGeneration.
So I changed that in the conversion script, such that it yields the right config. However, loading the final model via AutoModel still results in a T5Model even though the config now explicitly states the correct model type.
On the other hand, if I load via AutoModelForSeq2SeqLM it loads the lm_head. So, I guess that is a HF specific thing and not related to the conversion script per se.
Environment
To reproduce
Steps to reproduce the behavior:
hf_t5
modelscripts/inference/convert_composer_to_hf.py
Expected behavior
The model is saved as a HuggingFace snapshot without any issue
Additional context
Locally, I fixed this via simply loading with
AutoModel
and not viaAutoModelForCausalLM
.I guess this is fine.
The text was updated successfully, but these errors were encountered: