mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-06 05:32:50 +08:00
fix args
Former-commit-id: 65c5b0477c0e62691a1f8790670ba04d7f6d2804
This commit is contained in:
parent
16688b773a
commit
cae66bce3d
@ -479,7 +479,9 @@ python src/export_model.py \
|
|||||||
--adapter_name_or_path path_to_checkpoint \
|
--adapter_name_or_path path_to_checkpoint \
|
||||||
--template default \
|
--template default \
|
||||||
--finetuning_type lora \
|
--finetuning_type lora \
|
||||||
--export_dir path_to_export
|
--export_dir path_to_export \
|
||||||
|
--export_size 2 \
|
||||||
|
--export_legacy_format False
|
||||||
```
|
```
|
||||||
|
|
||||||
> [!WARNING]
|
> [!WARNING]
|
||||||
|
@ -479,7 +479,9 @@ python src/export_model.py \
|
|||||||
--adapter_name_or_path path_to_checkpoint \
|
--adapter_name_or_path path_to_checkpoint \
|
||||||
--template default \
|
--template default \
|
||||||
--finetuning_type lora \
|
--finetuning_type lora \
|
||||||
--export_dir path_to_export
|
--export_dir path_to_export \
|
||||||
|
--export_size 2 \
|
||||||
|
--export_legacy_format False
|
||||||
```
|
```
|
||||||
|
|
||||||
> [!WARNING]
|
> [!WARNING]
|
||||||
|
@ -102,7 +102,7 @@ class ModelArguments:
|
|||||||
default=1024,
|
default=1024,
|
||||||
metadata={"help": "The maximum length of the model inputs used for quantization."}
|
metadata={"help": "The maximum length of the model inputs used for quantization."}
|
||||||
)
|
)
|
||||||
export_lecacy_format: Optional[bool] = field(
|
export_legacy_format: Optional[bool] = field(
|
||||||
default=False,
|
default=False,
|
||||||
metadata={"help": "Whether or not to save the `.bin` files instead of `.safetensors`."}
|
metadata={"help": "Whether or not to save the `.bin` files instead of `.safetensors`."}
|
||||||
)
|
)
|
||||||
|
@ -50,7 +50,7 @@ def export_model(args: Optional[Dict[str, Any]] = None):
|
|||||||
model.save_pretrained(
|
model.save_pretrained(
|
||||||
save_directory=model_args.export_dir,
|
save_directory=model_args.export_dir,
|
||||||
max_shard_size="{}GB".format(model_args.export_size),
|
max_shard_size="{}GB".format(model_args.export_size),
|
||||||
safe_serialization=(not model_args.export_lecacy_format)
|
safe_serialization=(not model_args.export_legacy_format)
|
||||||
)
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user