mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 20:52:59 +08:00
fix export format
Former-commit-id: e165354facf7e69f535f9b7d99438f03dbf0293d
This commit is contained in:
parent
d0946f08db
commit
16688b773a
@ -758,6 +758,26 @@ register_template(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
register_template(
|
||||||
|
name="yuan",
|
||||||
|
prefix=[
|
||||||
|
"{{system}}"
|
||||||
|
],
|
||||||
|
prompt=[
|
||||||
|
"{{query}}",
|
||||||
|
{"token": "<sep>"}
|
||||||
|
],
|
||||||
|
system="",
|
||||||
|
sep=[
|
||||||
|
"\n"
|
||||||
|
],
|
||||||
|
stop_words=[
|
||||||
|
"<eod>"
|
||||||
|
],
|
||||||
|
replace_eos=True
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
register_template(
|
register_template(
|
||||||
name="zephyr",
|
name="zephyr",
|
||||||
prefix=[
|
prefix=[
|
||||||
|
@ -102,6 +102,10 @@ class ModelArguments:
|
|||||||
default=1024,
|
default=1024,
|
||||||
metadata={"help": "The maximum length of the model inputs used for quantization."}
|
metadata={"help": "The maximum length of the model inputs used for quantization."}
|
||||||
)
|
)
|
||||||
|
export_lecacy_format: Optional[bool] = field(
|
||||||
|
default=False,
|
||||||
|
metadata={"help": "Whether or not to save the `.bin` files instead of `.safetensors`."}
|
||||||
|
)
|
||||||
|
|
||||||
def __post_init__(self):
|
def __post_init__(self):
|
||||||
self.compute_dtype = None
|
self.compute_dtype = None
|
||||||
|
@ -47,7 +47,11 @@ def export_model(args: Optional[Dict[str, Any]] = None):
|
|||||||
|
|
||||||
model.config.use_cache = True
|
model.config.use_cache = True
|
||||||
model = model.to("cpu")
|
model = model.to("cpu")
|
||||||
model.save_pretrained(model_args.export_dir, max_shard_size="{}GB".format(model_args.export_size))
|
model.save_pretrained(
|
||||||
|
save_directory=model_args.export_dir,
|
||||||
|
max_shard_size="{}GB".format(model_args.export_size),
|
||||||
|
safe_serialization=(not model_args.export_lecacy_format)
|
||||||
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
tokenizer.padding_side = "left" # restore padding side
|
tokenizer.padding_side = "left" # restore padding side
|
||||||
|
Loading…
x
Reference in New Issue
Block a user