Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BFCL] Evaluation with Correct Precision Settings for Locally-Hosted Models #575

Closed
HuanzhiMao opened this issue Aug 10, 2024 · 1 comment · Fixed by #580
Closed

[BFCL] Evaluation with Correct Precision Settings for Locally-Hosted Models #575

HuanzhiMao opened this issue Aug 10, 2024 · 1 comment · Fixed by #580
Labels
BFCL-General General BFCL Issue

Comments

@HuanzhiMao
Copy link
Collaborator

The following models are intended to be evaluated using bfloat16 precision instead of float16 according to their model card on HuggingFace. We should change the default precision setting for their model handlers. This means they cannot be evaluated using v100 GPUs.

  • deepseek-ai/deepseek-coder-6.7b-instruct
  • google/gemma-7b-it
  • meetkai/functionary-small-v2.2-FC
  • meetkai/functionary-medium-v2.2-FC
  • meetkai/functionary-small-v2.4-FC
  • meetkai/functionary-medium-v2.4-FC
  • NousResearch/Hermes-2-Pro-Llama-3-70B
  • NousResearch/Hermes-2-Pro-Mistral-7B
  • NousResearch/Hermes-2-Theta-Llama-3-8B
  • NousResearch/Hermes-2-Theta-Llama-3-70B
  • meta-llama/Meta-Llama-3-8B-Instruct
  • meta-llama/Meta-Llama-3-70B-Instruct
  • ibm-granite/granite-20b-functioncalling
  • THUDM/glm-4-9b-chat
@HuanzhiMao HuanzhiMao added the BFCL-General General BFCL Issue label Aug 10, 2024
@HuanzhiMao
Copy link
Collaborator Author

HuanzhiMao commented Aug 10, 2024

Starting with #572:
We will have all the locally-hosted models' responses generated with bloat16 using A6000 GPU.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
BFCL-General General BFCL Issue
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant