You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've been able to convert files from HF format to f16 and 4bit, but I've not been able to figure out what config.json (or what changes to the config.json) to use when attempting to evaluate 4 bit quantized models. (Which is needed when trying to evaluate models)
I've tried to just change the torch_dtype to int4, but that doesn't seem to work.
For context, this is that the current config.json for llama 13b looks like:
Update: turns out lm evals does not support 4 bit quantizes yet, will need to work on that first (or if anyone wants to work on it, please see: EleutherAI/lm-evaluation-harness#417
I've been able to convert files from HF format to f16 and 4bit, but I've not been able to figure out what config.json (or what changes to the config.json) to use when attempting to evaluate 4 bit quantized models. (Which is needed when trying to evaluate models)
I've tried to just change the torch_dtype to int4, but that doesn't seem to work.
For context, this is that the current config.json for llama 13b looks like:
I would really appreciate any help. Thank you!
The text was updated successfully, but these errors were encountered: