Skip to content

Commit

Permalink
llm_layer_config_bugfix (#406)
Browse files Browse the repository at this point in the history
Signed-off-by: Zhang, Weiwei1 <[email protected]>
Co-authored-by: Wenhua Cheng <[email protected]>
  • Loading branch information
WeiweiZhang1 and wenhuach21 authored Jan 7, 2025
1 parent bf8d68d commit 7714796
Showing 1 changed file with 0 additions and 1 deletion.
1 change: 0 additions & 1 deletion auto_round/script/llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -400,7 +400,6 @@ def tune(args):
f"{n} will not be quantized due to its shape not being divisible by 32,"
" resulting in an exporting issue to autogptq")

layer_config = {}
not_quantize_layer_names = get_fp_layer_names(model, args.fp_layers)
for name in not_quantize_layer_names:
layer_config[name] = {"bits": 16}
Expand Down

0 comments on commit 7714796

Please sign in to comment.