Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

refactor TurbomindModelConfig #2364

Merged
merged 34 commits into from
Sep 2, 2024
Merged

refactor TurbomindModelConfig #2364

merged 34 commits into from
Sep 2, 2024

Conversation

lvhan028
Copy link
Collaborator

No description provided.

@lvhan028 lvhan028 requested a review from lzhangzz August 28, 2024 08:44
@lvhan028 lvhan028 changed the title Replace config.ini by config.yaml refactor TurbomindModelConfig Aug 28, 2024
Comment on lines +231 to +232
group_size = _group_size

Copy link
Collaborator

@irexyc irexyc Sep 2, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Shoud we assert _group_size ==128?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes.

# Compatible to awq models that are quantized by lmdeploy (<=v0.3.0)
    if not group_size:
        group_size = 128

    if engine_config.model_format in ['awq', 'gptq']:
        assert group_size == 128, \
            f'model format is "{engine_config.model_format}" ' \
            f'but group_size is {group_size}. Currently, only 128 ' \
            'is supported'

@irexyc
Copy link
Collaborator

irexyc commented Sep 2, 2024

test /mnt/140/InternLM/internlm2_5-7b-chat convert, load from workspace and load from hf.

@lvhan028 lvhan028 merged commit f4ee599 into InternLM:main Sep 2, 2024
9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants