-
Notifications
You must be signed in to change notification settings - Fork 169
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove is_gpt_fast flag #172
Conversation
Summary: It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now. in the future we'll refactor everything to use tensor subclass. Test Plan: tested locally to make sure `test_8da4w_quantizer_eval` still works Reviewers: Subscribers: Tasks: Tags:
@@ -268,57 +268,6 @@ def test_8da4w_quantizer_eval(self): | |||
f"accuracy regressed from 8.23 to {result['results']['wikitext']['word_perplexity,none']}" | |||
) | |||
|
|||
@unittest.skip("skipping until we get checkpoints for gpt-fast") | |||
def test_gptq_quantizer_gpt_fast(self): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is this test not useful to keep around or some new version of it? Or more generally can we no longer use gpt-fast and ao together?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
oh next test tests gpt fast code path: test_gptq_quantizer_int4wo, this was initially added because we were trying to merge the gpt fast code path and 8da4w code path together in the same quantizer code path and use a flag to distinguish them, but now we just duplicated the quantizer code
so since we removed the gpt-fast code path in Int8DynActInt4WeightGPTQQuantizer, we no longer need to test this
Summary: It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now. in the future we'll refactor everything to use tensor subclass. Test Plan: tested locally to make sure `test_8da4w_quantizer_eval` still works Reviewers: Subscribers: Tasks: Tags:
Summary:
It was added before to merge the code for 8da4w and int4 weight only quant, but later we just duplicated the quantizer code, so we can safely remove this now.
in the future we'll refactor everything to use tensor subclass.
Test Plan:
tested locally to make sure
test_8da4w_quantizer_eval
still worksReviewers:
Subscribers:
Tasks:
Tags: