diff --git a/src/llamafactory/data/loader.py b/src/llamafactory/data/loader.py index 0849b603..385909fc 100644 --- a/src/llamafactory/data/loader.py +++ b/src/llamafactory/data/loader.py @@ -96,7 +96,7 @@ def _load_single_dataset( data_files=data_files, split=dataset_attr.split, cache_dir=cache_dir, - token=model_args.ms_hub_token, + token=model_args.om_hub_token, streaming=(data_args.streaming and (dataset_attr.load_from != "file")), ) except ImportError: diff --git a/src/llamafactory/extras/constants.py b/src/llamafactory/extras/constants.py index bf2386f4..f0f6d619 100644 --- a/src/llamafactory/extras/constants.py +++ b/src/llamafactory/extras/constants.py @@ -563,7 +563,7 @@ register_model_group( "Gemma-2-9B-Instruct": { DownloadSource.DEFAULT: "google/gemma-2-9b-it", DownloadSource.MODELSCOPE: "LLM-Research/gemma-2-9b-it", - DownloadSource.MODELERS: "LlamaFactory/Qwen2-VL-2B-Instruct" + DownloadSource.MODELERS: "LlamaFactory/gemma-2-2b-it" }, "Gemma-2-27B-Instruct": { DownloadSource.DEFAULT: "google/gemma-2-27b-it", diff --git a/src/llamafactory/hparams/model_args.py b/src/llamafactory/hparams/model_args.py index cd2f1867..9847d707 100644 --- a/src/llamafactory/hparams/model_args.py +++ b/src/llamafactory/hparams/model_args.py @@ -267,6 +267,10 @@ class ModelArguments(QuantizationArguments, ProcessorArguments, ExportArguments, default=None, metadata={"help": "Auth token to log in with ModelScope Hub."}, ) + om_hub_token: Optional[str] = field( + default=None, + metadata={"help": "Auth token to log in with Modelers Hub."}, + ) print_param_status: bool = field( default=False, metadata={"help": "For debugging purposes, print the status of the parameters in the model."},