mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 00:28:10 +08:00
update
Former-commit-id: 516ed0ea5fed8c74fe3669a7e85dd89b5a0ec3c2
This commit is contained in:
parent
a20c2b6ecf
commit
fca814b30d
@ -66,10 +66,10 @@ def run_dpo(
|
|||||||
# Update arguments
|
# Update arguments
|
||||||
training_args.remove_unused_columns = False # important for multimodal and pairwise dataset
|
training_args.remove_unused_columns = False # important for multimodal and pairwise dataset
|
||||||
|
|
||||||
effi_token_num = 0.0
|
effective_token_num = 0.0
|
||||||
for data in dataset_module["train_dataset"]:
|
for data in dataset_module["train_dataset"]:
|
||||||
effi_token_num += len(data["chosen_input_ids"])
|
effective_token_num += len(data["chosen_input_ids"])
|
||||||
effi_token_num += len(data["rejected_input_ids"])
|
effective_token_num += len(data["rejected_input_ids"])
|
||||||
|
|
||||||
# Initialize our Trainer
|
# Initialize our Trainer
|
||||||
trainer = CustomDPOTrainer(
|
trainer = CustomDPOTrainer(
|
||||||
@ -87,7 +87,7 @@ def run_dpo(
|
|||||||
if training_args.do_train:
|
if training_args.do_train:
|
||||||
train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint)
|
train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint)
|
||||||
train_result.metrics["effective_tokens_per_sec"] = (
|
train_result.metrics["effective_tokens_per_sec"] = (
|
||||||
effi_token_num * train_result.metrics["epoch"] / train_result.metrics["train_runtime"]
|
effective_token_num * train_result.metrics["epoch"] / train_result.metrics["train_runtime"]
|
||||||
)
|
)
|
||||||
if dist.is_initialized():
|
if dist.is_initialized():
|
||||||
train_result.metrics["effective_tokens_per_sec"] = (
|
train_result.metrics["effective_tokens_per_sec"] = (
|
||||||
|
Loading…
x
Reference in New Issue
Block a user