From 79a09c8fba916862613f574a4d6e808902e764a5 Mon Sep 17 00:00:00 2001 From: Qubitium-ModelCloud Date: Wed, 3 Jul 2024 00:00:09 +0800 Subject: [PATCH] Revert "fix opt fc1/fc2 layer modules should not be quantized (#118)" (#149) This reverts commit 5bf289a1466b1fd9e431b65262fc5b2ec4e0a158. --- gptqmodel/models/opt.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/gptqmodel/models/opt.py b/gptqmodel/models/opt.py index c2e27c05..1cbf407e 100644 --- a/gptqmodel/models/opt.py +++ b/gptqmodel/models/opt.py @@ -15,6 +15,6 @@ class OPTGPTQ(BaseGPTQModel): layer_modules = [ ["self_attn.k_proj", "self_attn.v_proj", "self_attn.q_proj"], ["self_attn.out_proj"], - # ["fc1"], disabled: not a good candidate for quantization - # ["fc2"], disabled: not a good candidate for quantization + ["fc1"], + ["fc2"], ]