-
Notifications
You must be signed in to change notification settings - Fork 14.6k
Detect GigaChat3-10-A1.8B as deepseek lite #17420
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
I added a test Q8_0 gguf here: https://huggingface.co/ubergarm/GigaChat3-10B-A1.8B-GGUF/tree/main |
|
Oops i need to get rid of an accidently commit file, will force push to fix. |
Hardcodes checking number of layers to detect if lite version of deepseek.
3cddaab to
e1bfe51
Compare
|
Perplexity seems reasonable:
Always a bit funky when quants have lower perplexity than the original bf16... though it happens sometimes and in this case the values are very similar within the noise. |
CISC
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you add a comment describing which models we are detecting in case this needs to be fine-tuned in the future?
deepseek lite variants include DeepSeek-V2-Lite, GigaChat3-10B-A1.8B
|
Thank you! |
* Detect GigaChat3-10-A1.8B as deepseek lite Hardcodes checking number of layers to detect if lite version of deepseek. * Add commnent identifying deepseek lite variants deepseek lite variants include DeepSeek-V2-Lite, GigaChat3-10B-A1.8B
Hardcodes checking number of layers to detect if a model is the lite version of deepseek.
Tested with bf16 and q8_0 version of GigaChat3-10B-A1.8B and discussed realizing it was a
liteversion similar to DeepSeek-V2-Lite. That model had 27 layers, but GigaChat3 has 26 and that is used to detect thelitevariant as discussed here: https://huggingface.co/ai-sage/GigaChat3-10B-A1.8B/discussions/1#691fb161ac024c8eb626ab36I'd like if anyone else could test. I'll update after testing perplexity to make sure the value looks sane. I haven't uploaded a gguf yet as the template has a parse error and wanted to get it updated before baking it in. That is discussed here: https://huggingface.co/ai-sage/GigaChat3-702B-A36B-preview-bf16/discussions/1