mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-02 03:32:50 +08:00
tiny fix
Former-commit-id: b8a034807e97730bc95a5f0bf0ca0763d7c6824c
This commit is contained in:
parent
6ab22a0181
commit
36790c4e32
@ -4,15 +4,15 @@
|
|||||||
|
|
||||||
|
|
||||||
from transformers import HfArgumentParser, TrainingArguments
|
from transformers import HfArgumentParser, TrainingArguments
|
||||||
from utils import ModelArguments, load_pretrained
|
from utils import ModelArguments, FinetuningArguments, load_pretrained
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
parser = HfArgumentParser((ModelArguments, TrainingArguments))
|
parser = HfArgumentParser((ModelArguments, TrainingArguments, FinetuningArguments))
|
||||||
model_args, training_args = parser.parse_args_into_dataclasses()
|
model_args, training_args, finetuning_args = parser.parse_args_into_dataclasses()
|
||||||
|
|
||||||
model, tokenizer = load_pretrained(model_args)
|
model, tokenizer = load_pretrained(model_args, finetuning_args)
|
||||||
model.save_pretrained(training_args.output_dir, max_shard_size="1GB")
|
model.save_pretrained(training_args.output_dir, max_shard_size="1GB")
|
||||||
tokenizer.save_pretrained(training_args.output_dir)
|
tokenizer.save_pretrained(training_args.output_dir)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user