diff --git a/examples/extras/badam/llama3_lora_sft.yaml b/examples/extras/badam/llama3_lora_sft.yaml index 242e63ab..a78de2fa 100644 --- a/examples/extras/badam/llama3_lora_sft.yaml +++ b/examples/extras/badam/llama3_lora_sft.yaml @@ -37,5 +37,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/fsdp_qlora/llama3_lora_sft.yaml b/examples/extras/fsdp_qlora/llama3_lora_sft.yaml index 920d8fdb..348459b8 100644 --- a/examples/extras/fsdp_qlora/llama3_lora_sft.yaml +++ b/examples/extras/fsdp_qlora/llama3_lora_sft.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/galore/llama3_full_sft.yaml b/examples/extras/galore/llama3_full_sft.yaml index 3db31fed..605545de 100644 --- a/examples/extras/galore/llama3_full_sft.yaml +++ b/examples/extras/galore/llama3_full_sft.yaml @@ -38,5 +38,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/llama_pro/llama3_freeze_sft.yaml b/examples/extras/llama_pro/llama3_freeze_sft.yaml index 214f411a..444a1113 100644 --- a/examples/extras/llama_pro/llama3_freeze_sft.yaml +++ b/examples/extras/llama_pro/llama3_freeze_sft.yaml @@ -36,5 +36,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/loraplus/llama3_lora_sft.yaml b/examples/extras/loraplus/llama3_lora_sft.yaml index 9936bcd3..960f613e 100644 --- a/examples/extras/loraplus/llama3_lora_sft.yaml +++ b/examples/extras/loraplus/llama3_lora_sft.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/mod/llama3_full_sft.yaml b/examples/extras/mod/llama3_full_sft.yaml index edfec44e..df03c1e0 100644 --- a/examples/extras/mod/llama3_full_sft.yaml +++ b/examples/extras/mod/llama3_full_sft.yaml @@ -35,5 +35,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/full_multi_gpu/llama3_full_sft.yaml b/examples/full_multi_gpu/llama3_full_sft.yaml index b8873e3a..40b62f24 100644 --- a/examples/full_multi_gpu/llama3_full_sft.yaml +++ b/examples/full_multi_gpu/llama3_full_sft.yaml @@ -37,5 +37,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_gpu/llama3_lora_sft.yaml b/examples/lora_multi_gpu/llama3_lora_sft.yaml index 5e5dd9e6..9be3c780 100644 --- a/examples/lora_multi_gpu/llama3_lora_sft.yaml +++ b/examples/lora_multi_gpu/llama3_lora_sft.yaml @@ -37,5 +37,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml b/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml index e8dee216..41152243 100644 --- a/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml +++ b/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_npu/llama3_lora_sft_ds.yaml b/examples/lora_multi_npu/llama3_lora_sft_ds.yaml index 825b8450..1ed24d04 100644 --- a/examples/lora_multi_npu/llama3_lora_sft_ds.yaml +++ b/examples/lora_multi_npu/llama3_lora_sft_ds.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_dpo.yaml b/examples/lora_single_gpu/llama3_lora_dpo.yaml index 62752e57..158c9e04 100644 --- a/examples/lora_single_gpu/llama3_lora_dpo.yaml +++ b/examples/lora_single_gpu/llama3_lora_dpo.yaml @@ -36,5 +36,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_kto.yaml b/examples/lora_single_gpu/llama3_lora_kto.yaml index 6f689818..ead221e9 100644 --- a/examples/lora_single_gpu/llama3_lora_kto.yaml +++ b/examples/lora_single_gpu/llama3_lora_kto.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_pretrain.yaml b/examples/lora_single_gpu/llama3_lora_pretrain.yaml index 54c5d89a..9167a893 100644 --- a/examples/lora_single_gpu/llama3_lora_pretrain.yaml +++ b/examples/lora_single_gpu/llama3_lora_pretrain.yaml @@ -33,5 +33,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_reward.yaml b/examples/lora_single_gpu/llama3_lora_reward.yaml index c82f9414..91663057 100644 --- a/examples/lora_single_gpu/llama3_lora_reward.yaml +++ b/examples/lora_single_gpu/llama3_lora_reward.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_sft.yaml b/examples/lora_single_gpu/llama3_lora_sft.yaml index 429cb6af..cc93d05a 100644 --- a/examples/lora_single_gpu/llama3_lora_sft.yaml +++ b/examples/lora_single_gpu/llama3_lora_sft.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llava1_5_lora_sft.yaml b/examples/lora_single_gpu/llava1_5_lora_sft.yaml index acab4884..95c1d40d 100644 --- a/examples/lora_single_gpu/llava1_5_lora_sft.yaml +++ b/examples/lora_single_gpu/llava1_5_lora_sft.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml b/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml index 53cc12e2..23301de5 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml b/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml index 1a92f822..40a290a3 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml b/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml index c7f72c66..6652d8cf 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml b/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml index 45caf17c..323ea7c6 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/src/llamafactory/extras/env.py b/src/llamafactory/extras/env.py index 27453a6b..059730f1 100644 --- a/src/llamafactory/extras/env.py +++ b/src/llamafactory/extras/env.py @@ -51,4 +51,4 @@ def print_env() -> None: info["vLLM version"] = vllm.__version__ - print("\n".join(["- {}: {}".format(key, value) for key, value in info.items()]) + "\n") + print("\n" + "\n".join(["- {}: {}".format(key, value) for key, value in info.items()]) + "\n") diff --git a/src/llamafactory/webui/runner.py b/src/llamafactory/webui/runner.py index 6e1facef..d35fd903 100644 --- a/src/llamafactory/webui/runner.py +++ b/src/llamafactory/webui/runner.py @@ -200,7 +200,7 @@ class Runner: # eval config if get("train.val_size") > 1e-6 and args["stage"] != "ppo": args["val_size"] = get("train.val_size") - args["evaluation_strategy"] = "steps" + args["eval_strategy"] = "steps" args["eval_steps"] = args["save_steps"] args["per_device_eval_batch_size"] = args["per_device_train_batch_size"]