-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support different prompting templates #27
Comments
I would agree with a need for this functionality. On several models, Amica tends to trend toward ending all responses with "agent: {... something they *think the user would say}", generating a potential user response instead of waiting for the next response from the actual user. This then tends to bias future generation toward the Amica 'user' prompt tokens instead of the actual following user reply. |
here's a diff that switches to the chatml format, just for other users who might find this issue |
Arbius has a $150 AIUS bounty for this issue! Brief: Fix/Complete Issue, suggest adding it as expandable under system prompt page. Please read carefully: To begin work on a bounty, reply by saying “I claim this bounty” - you will have 48 hours to submit your PR before someone else may attempt to claim this bounty. To complete the bounty, within 48 hours of claiming, reply with a link to your PR referencing this issue and an Ethereum address. You must comply with reviewers comments and have the PR merged to receive the bounty reward. Please be sure to focus on quality submissions to minimize the amount of time reviewers must take. |
Thank you for Amica, I'm having a lot of fun with it!
I tried using the yi-34B models with llama.cpp with it, and generally it works, but the results aren't optimal. For example, the reply always ends with "<|im_end|>" which ends up visible in the conversation.
It would be great to have a way to not just change the system prompt, but to change the prompting template altogether. Here's ChatML, which Yi is trained to use:
The text was updated successfully, but these errors were encountered: