mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-01 11:12:50 +08:00
[tracking] fix swanlab hparams (#8532)
Co-authored-by: Yaowei Zheng <hiyouga@buaa.edu.cn>
This commit is contained in:
parent
e117e3c2b7
commit
8e7727f4ee
@ -164,6 +164,9 @@ def _check_extra_dependencies(
|
|||||||
if finetuning_args.use_adam_mini:
|
if finetuning_args.use_adam_mini:
|
||||||
check_version("adam-mini", mandatory=True)
|
check_version("adam-mini", mandatory=True)
|
||||||
|
|
||||||
|
if finetuning_args.use_swanlab:
|
||||||
|
check_version("swanlab", mandatory=True)
|
||||||
|
|
||||||
if finetuning_args.plot_loss:
|
if finetuning_args.plot_loss:
|
||||||
check_version("matplotlib", mandatory=True)
|
check_version("matplotlib", mandatory=True)
|
||||||
|
|
||||||
@ -345,6 +348,9 @@ def get_train_args(args: Optional[Union[dict[str, Any], list[str]]] = None) -> _
|
|||||||
if finetuning_args.finetuning_type == "lora":
|
if finetuning_args.finetuning_type == "lora":
|
||||||
# https://github.com/huggingface/transformers/blob/v4.50.0/src/transformers/trainer.py#L782
|
# https://github.com/huggingface/transformers/blob/v4.50.0/src/transformers/trainer.py#L782
|
||||||
training_args.label_names = training_args.label_names or ["labels"]
|
training_args.label_names = training_args.label_names or ["labels"]
|
||||||
|
|
||||||
|
if "swanlab" in training_args.report_to and finetuning_args.use_swanlab:
|
||||||
|
training_args.report_to.remove("swanlab")
|
||||||
|
|
||||||
if (
|
if (
|
||||||
training_args.parallel_mode == ParallelMode.DISTRIBUTED
|
training_args.parallel_mode == ParallelMode.DISTRIBUTED
|
||||||
|
Loading…
x
Reference in New Issue
Block a user