Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove is_gpt_fast flag #172

Merged
merged 3 commits into from
Apr 24, 2024
Merged

Remove is_gpt_fast flag #172

merged 3 commits into from
Apr 24, 2024

Conversation

jerryzh168
Copy link
Contributor

Summary:
It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now.

in the future we'll refactor everything to use tensor subclass.

Test Plan:
tested locally to make sure test_8da4w_quantizer_eval still works

Reviewers:

Subscribers:

Tasks:

Tags:

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Apr 24, 2024
Summary:
It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now.

in the future we'll refactor everything to use tensor subclass.

Test Plan:
tested locally to make sure `test_8da4w_quantizer_eval` still works

Reviewers:

Subscribers:

Tasks:

Tags:
@@ -268,57 +268,6 @@ def test_8da4w_quantizer_eval(self):
f"accuracy regressed from 8.23 to {result['results']['wikitext']['word_perplexity,none']}"
)

@unittest.skip("skipping until we get checkpoints for gpt-fast")
def test_gptq_quantizer_gpt_fast(self):
Copy link
Member

@msaroufim msaroufim Apr 24, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

is this test not useful to keep around or some new version of it? Or more generally can we no longer use gpt-fast and ao together?

Copy link
Contributor Author

@jerryzh168 jerryzh168 Apr 24, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

oh next test tests gpt fast code path: test_gptq_quantizer_int4wo, this was initially added because we were trying to merge the gpt fast code path and 8da4w code path together in the same quantizer code path and use a flag to distinguish them, but now we just duplicated the quantizer code

so since we removed the gpt-fast code path in Int8DynActInt4WeightGPTQQuantizer, we no longer need to test this

@msaroufim msaroufim self-requested a review April 24, 2024 22:05
@msaroufim msaroufim merged commit 7fc5e26 into pytorch:main Apr 24, 2024
13 checks passed
dbyoung18 pushed a commit to dbyoung18/ao that referenced this pull request Jul 31, 2024
Summary:
It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now.

in the future we'll refactor everything to use tensor subclass.

Test Plan:
tested locally to make sure `test_8da4w_quantizer_eval` still works

Reviewers:

Subscribers:

Tasks:

Tags:
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants