Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Adapt to new quantization_config #38

Merged
merged 5 commits into from
Feb 17, 2025

Conversation

Angazenn
Copy link

@Angazenn Angazenn commented Feb 11, 2025

This pr mainly focuses on:

  1. adapting to new quantization_config generated from msmodelslim.
  2. removing unnecessary imports.
  3. disable warning when loading BasevLLMParameter

angazenn added 2 commits February 11, 2025 15:38
Signed-off-by: angazenn <zengyanjia@huawei.com>
Signed-off-by: angazenn <zengyanjia@huawei.com>
@Angazenn Angazenn changed the title [Feature] Adapt to new config.json from msmodelslim [Feature] Adapt to new quantization_config Feb 12, 2025
import mindie_turbo
except:
import mindie_turbo # noqa: F401
except Exception:
Copy link
Collaborator

@ganyi1996ppo ganyi1996ppo Feb 13, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why Exception rather than ImportError?

angazenn added 3 commits February 13, 2025 18:55
Signed-off-by: angazenn <zengyanjia@huawei.com>
Signed-off-by: angazenn <zengyanjia@huawei.com>
@wangxiyuan wangxiyuan merged commit 0037d46 into vllm-project:v0.7.1-dev Feb 17, 2025
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants