-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[PlayGround] Long generating cannot be done #1349
Comments
For now, there is a hard limitation of 2048 input tokens and a maximum of 1920 output tokens. We might consider increasing these numbers in the future. |
If the output > 1920 tokens (or lager maximum amount in the future), is there any way to let it continue to output all? |
Is there are reason as to why context is static for Tabby? @wsxiaoys |
Hey - could you elaborate a bit? static context length is kind of a intrinsic thing to transformer based LLMs. |
Sorry. Dependent on the models, the context can be increased to a certain size. You state that there is a hard limitation of input and output tokens, is that hard coded in Tabby? Or variable due to the model being used? |
Ah - I got you point. It does make sense to read this value from either the registry or from gguf files directly. Filing #1402 to track |
Describe the bug
In playground, when the generating is long, it cannot finish generating in one time, in this case I typied "continue" to let it keep going. But it didn't work, it repeated from the beginning or returned empty.
Information about your version
tabby 0.7.0
Information about your GPU
Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered: