Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ENH: support more GPTQ and AWQ format for some models #1243

Merged
merged 11 commits into from
Apr 5, 2024
Merged

ENH: support more GPTQ and AWQ format for some models #1243

merged 11 commits into from
Apr 5, 2024

Conversation

xiaodouzi666
Copy link
Contributor

No description provided.

@XprobeBot XprobeBot added the enhancement New feature or request label Apr 3, 2024
@XprobeBot XprobeBot added this to the v0.10.1 milestone Apr 3, 2024
@xiaodouzi666 xiaodouzi666 reopened this Apr 3, 2024
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
xinference/model/llm/llm_family.json Show resolved Hide resolved
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
xinference/model/llm/llm_family.json Outdated Show resolved Hide resolved
@qinxuye
Copy link
Contributor

qinxuye commented Apr 3, 2024

Lack of llama-2-chat 70b awq, and 13b awq&gptq

@qinxuye qinxuye changed the title ENH: Modify llm_family.json ENH: support more GPTQ and AWQ format for some models Apr 3, 2024
Copy link
Contributor

@qinxuye qinxuye left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@qinxuye qinxuye merged commit 3b922b6 into xorbitsai:main Apr 5, 2024
10 of 12 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants