Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bugs when fine-tune llama-2-7b with instructions using llama-2's conversation template #2871

Closed
YJiangcm opened this issue Dec 29, 2023 · 0 comments · Fixed by #2996 or #3006
Closed

Comments

@YJiangcm
Copy link

Thanks for your great work! I met some problems when using fastchat/train/train.py to fine-tune a llama-2-7b by using llama-2's conversation template.

I have changed the get_conversation_template("vicuna") to get_conversation_template("llama-2"), and delated assert conv.sep_style == SeparatorStyle.ADD_COLON_TWO. However, the tokenization mismatch warning was reported, and the training loss was always 0.

WARNING: tokenization mismatch: 78 vs. 80.
#turn = 1. (ignored)

Could you please tell me how to adapt the code to llama-2's conversation template? Thanks a lot!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
1 participant