mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 20:52:59 +08:00
fix loading dtype
Former-commit-id: c874e764b8334c18091233c3781009a39d67e794
This commit is contained in:
parent
1faba4b704
commit
089785c71b
@ -88,11 +88,10 @@ def load_model_and_tokenizer(
|
|||||||
tokenizer._pad = MethodType(PreTrainedTokenizerBase._pad, tokenizer)
|
tokenizer._pad = MethodType(PreTrainedTokenizerBase._pad, tokenizer)
|
||||||
|
|
||||||
# Set model dtype
|
# Set model dtype
|
||||||
if model_args.compute_dtype is not None:
|
if model_args.compute_dtype is not None: # for training
|
||||||
setattr(config, "torch_dtype", model_args.compute_dtype)
|
setattr(config, "torch_dtype", model_args.compute_dtype)
|
||||||
else: # priority: bf16 > fp16 > fp32
|
else: # for evaluation, priority: bf16 > fp16 > fp32
|
||||||
optim_dtype = infer_optim_dtype(model_dtype=getattr(config, "torch_dtype", None))
|
model_args.compute_dtype = infer_optim_dtype(model_dtype=getattr(config, "torch_dtype", None))
|
||||||
setattr(config, "torch_dtype", optim_dtype)
|
|
||||||
|
|
||||||
# Fix config (for Qwen)
|
# Fix config (for Qwen)
|
||||||
if getattr(config, "model_type", None) == "qwen":
|
if getattr(config, "model_type", None) == "qwen":
|
||||||
@ -185,7 +184,7 @@ def load_model_and_tokenizer(
|
|||||||
model = AutoModelForCausalLM.from_pretrained(
|
model = AutoModelForCausalLM.from_pretrained(
|
||||||
model_to_load,
|
model_to_load,
|
||||||
config=config,
|
config=config,
|
||||||
torch_dtype=getattr(config, "torch_dtype"),
|
torch_dtype=model_args.compute_dtype,
|
||||||
low_cpu_mem_usage=(not is_deepspeed_zero3_enabled()),
|
low_cpu_mem_usage=(not is_deepspeed_zero3_enabled()),
|
||||||
**config_kwargs
|
**config_kwargs
|
||||||
)
|
)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user