mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-15 08:08:09 +08:00
tiny fix
Former-commit-id: 3b6e1132c4d203e6d5376cf97e81cc160697c822
This commit is contained in:
parent
48d2e6d7fe
commit
30855b924a
@ -169,11 +169,14 @@ def _configure_quantization(
|
|||||||
quantization_config["use_exllama"] = False # disable exllama
|
quantization_config["use_exllama"] = False # disable exllama
|
||||||
|
|
||||||
if quantization_config.get("quant_method", None) == "aqlm":
|
if quantization_config.get("quant_method", None) == "aqlm":
|
||||||
|
require_version(
|
||||||
|
"transformers>=4.39.0.dev0", "To fix: pip install git+https://github.com/huggingface/transformers.git"
|
||||||
|
)
|
||||||
quantization_config["bits"] = 2
|
quantization_config["bits"] = 2
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
"Loading {}-bit {}-quantized model.".format(
|
"Loading {}-bit {}-quantized model.".format(
|
||||||
quantization_config.get("bits", "?"), quantization_config.get("quant_method", None)
|
quantization_config.get("bits", "?"), str(quantization_config.get("quant_method", "")).upper()
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user