Skip to content

Commit

Permalink
Update GPTQ_Loader.py
Browse files Browse the repository at this point in the history
Correcting decoder layer for renamed class.
askmyteapot authored Mar 17, 2023

Verified

This commit was signed with the committer’s verified signature.
knqyf263 Teppei Fukuda
1 parent ee164d1 commit 53b6a66
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion modules/GPTQ_loader.py
Original file line number Diff line number Diff line change
@@ -61,7 +61,7 @@ def load_quantized(model_name):
max_memory[i] = f"{shared.args.gpu_memory[i]}GiB"
max_memory['cpu'] = f"{shared.args.cpu_memory or '99'}GiB"

device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LLaMADecoderLayer"])
device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LlamaDecoderLayer"])
model = accelerate.dispatch_model(model, device_map=device_map)

# Single GPU

0 comments on commit 53b6a66

Please sign in to comment.