From fb9280a0a7522b544c03c9e133bfa2787eda7655 Mon Sep 17 00:00:00 2001 From: hiyouga <467089858@qq.com> Date: Sun, 8 Sep 2024 23:43:35 +0800 Subject: [PATCH] release v0.9.0 Former-commit-id: 54b5c4b8195d23bd9dcc1921af9910d5bdd181fd --- src/llamafactory/extras/constants.py | 10 ++++++++++ src/llamafactory/extras/env.py | 2 +- 2 files changed, 11 insertions(+), 1 deletion(-) diff --git a/src/llamafactory/extras/constants.py b/src/llamafactory/extras/constants.py index 9c00cea9..d4f3fbbe 100644 --- a/src/llamafactory/extras/constants.py +++ b/src/llamafactory/extras/constants.py @@ -763,6 +763,7 @@ register_model_group( }, "LLaMA3.1-405B": { DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-405B", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-405B", }, "LLaMA3.1-8B-Chat": { DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-8B-Instruct", @@ -774,6 +775,7 @@ register_model_group( }, "LLaMA3.1-405B-Chat": { DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-405B-Instruct", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-405B-Instruct", }, }, template="llama3", @@ -1357,15 +1359,19 @@ register_model_group( }, "Qwen2VL-2B-int8-Chat": { DownloadSource.DEFAULT: "Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8", + DownloadSource.MODELSCOPE: "qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8", }, "Qwen2VL-2B-int4-Chat": { DownloadSource.DEFAULT: "Qwen/Qwen2-VL-2B-Instruct-AWQ", + DownloadSource.MODELSCOPE: "qwen/Qwen2-VL-2B-Instruct-AWQ", }, "Qwen2VL-7B-int8-Chat": { DownloadSource.DEFAULT: "Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8", + DownloadSource.MODELSCOPE: "qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8", }, "Qwen2VL-7B-int4-Chat": { DownloadSource.DEFAULT: "Qwen/Qwen2-VL-7B-Instruct-AWQ", + DownloadSource.MODELSCOPE: "qwen/Qwen2-VL-7B-Instruct-AWQ", }, }, template="qwen2_vl", @@ -1646,15 +1652,19 @@ register_model_group( }, "Yi-Coder-1.5B": { DownloadSource.DEFAULT: "01-ai/Yi-Coder-1.5B", + DownloadSource.MODELSCOPE: "01ai/Yi-Coder-1.5B", }, "Yi-Coder-9B": { DownloadSource.DEFAULT: "01-ai/Yi-Coder-9B", + DownloadSource.MODELSCOPE: "01ai/Yi-Coder-9B", }, "Yi-Coder-1.5B-Chat": { DownloadSource.DEFAULT: "01-ai/Yi-Coder-1.5B-Chat", + DownloadSource.MODELSCOPE: "01ai/Yi-Coder-1.5B-Chat", }, "Yi-Coder-9B-Chat": { DownloadSource.DEFAULT: "01-ai/Yi-Coder-9B-Chat", + DownloadSource.MODELSCOPE: "01ai/Yi-Coder-9B-Chat", }, }, template="yi", diff --git a/src/llamafactory/extras/env.py b/src/llamafactory/extras/env.py index 5a2641de..741defea 100644 --- a/src/llamafactory/extras/env.py +++ b/src/llamafactory/extras/env.py @@ -26,7 +26,7 @@ import trl from transformers.utils import is_torch_cuda_available, is_torch_npu_available -VERSION = "0.8.4.dev0" +VERSION = "0.9.0" def print_env() -> None: