mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-16 20:00:36 +08:00
use low_cpu_mem_usage to speed up loading
This commit is contained in:
@@ -13,7 +13,7 @@ def main():
|
||||
model_args, training_args, finetuning_args = parser.parse_args_into_dataclasses()
|
||||
|
||||
model, tokenizer = load_pretrained(model_args, finetuning_args)
|
||||
model.save_pretrained(training_args.output_dir, max_shard_size="1GB")
|
||||
model.save_pretrained(training_args.output_dir, max_shard_size="10GB")
|
||||
tokenizer.save_pretrained(training_args.output_dir)
|
||||
|
||||
print("model and tokenizer have been saved at:", training_args.output_dir)
|
||||
|
||||
Reference in New Issue
Block a user