mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 06:12:50 +08:00
Merge pull request #4950 from liuwwang/main and fi
fix: Repair the issue where quantization failed after merging the adapter. Former-commit-id: 9839c6d1f6fd69287afeb3b6528571cf2bedf91a
This commit is contained in:
commit
29d9a9827a
@ -66,7 +66,7 @@ def save_model(
|
||||
error = ALERTS["err_no_dataset"][lang]
|
||||
elif export_quantization_bit not in GPTQ_BITS and not checkpoint_path:
|
||||
error = ALERTS["err_no_adapter"][lang]
|
||||
elif export_quantization_bit in GPTQ_BITS and isinstance(checkpoint_path, list):
|
||||
elif export_quantization_bit in GPTQ_BITS and checkpoint_path and isinstance(checkpoint_path, list):
|
||||
error = ALERTS["err_gptq_lora"][lang]
|
||||
|
||||
if error:
|
||||
|
Loading…
x
Reference in New Issue
Block a user