Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Generating output from finetuned model using LLM.generate() method vs litgpt chat cli command #1937

Open
asmitnayak opened this issue Feb 16, 2025 · 0 comments
Labels
question Further information is requested

Comments

@asmitnayak
Copy link

asmitnayak commented Feb 16, 2025

I performed lora finetuning on Qwen2.5-3B-Instruct model on my custom dataset, and it seems that when I generate content using the litgpt chat cli interface the finetuned model performs flawlessly. However, when using it through the python api, following the tutorial the model outputs are random characters and basically rubbish. I tried using litserve and converting it into HuggingFace models but none worked.

Does anyone have any idea how to fix this.

P.S. Currently I am using a custom implementation of litgpt\chat\base.py to get my code working but this is pretty slow and I wish to parallelize the generation.

Thanks for any help.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant