You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Using qwen-2.5-coder-7b-instruct, Jan allows a maximum context length of 2048 tokens. However, both according to qwen's website as well as llama.cpp's output (most recent version from github at the time of writing), the maximum context length is 131072 tokens.
From (janhq/jan#3714 (comment)),
we know that the context length for GGUF models are 32768.
The full context length of 131072, one has to refer to non-GGUF models.
Jan version
v0.5.4
Describe the Bug
Using qwen-2.5-coder-7b-instruct, Jan allows a maximum context length of 2048 tokens. However, both according to qwen's website as well as llama.cpp's output (most recent version from github at the time of writing), the maximum context length is 131072 tokens.
Unfortunately this makes the model within Jan not really usable for coding related tasks.
Steps to Reproduce
To my understanding, even without ROPE 2048 is way below the trained context length of qwen-2.5 coder.
Screenshots / Logs
Unfortunately, there are no logs under
~/jan/logs
(I did a find over all of my home folder for anapp.log
file, but none were found).Maybe I'll checkout the repo and debug this myself later on.
What is your OS?
The text was updated successfully, but these errors were encountered: