-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Causal models only supported for text-generation task, not summarization task #972
base: main
Are you sure you want to change the base?
Causal models only supported for text-generation task, not summarization task #972
Conversation
max_length = self._pipeline.model.config.max_position_embeddings | ||
# If the model is of the HF Hub the odds of this being wrong are low, but it's still good to check that the | ||
# tokenizer model and the model have the same max_position_embeddings | ||
if self._pipeline.tokenizer.model_max_length != max_length: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Because some models (esp older ones) Don't have a tokenizer_config file! Looking at you, facebook/bart-large-cnn lol https://huggingface.co/facebook/bart-large-cnn/discussions/71
It was pointed out that I accidentally remove text-generation support when merging LiteLLM. I'll put back that code #978 and then refactor this PR based on that code. Thanks! |
…model-not-supported-for-summarization
What's changing
Refs #970
How to test it
CI Tests, also load the Lumigator UI and make sure you can still test the BART CNN model.
Additional notes for reviewers
I already...
/docs
)