diff --git a/README.md b/README.md index 89c1ccb9..ab1e44f1 100644 --- a/README.md +++ b/README.md @@ -164,7 +164,7 @@ Compared to ChatGLM's [P-Tuning](https://github.com/THUDM/ChatGLM2-6B/tree/main/ | [InternLM2](https://huggingface.co/internlm) | 7B/20B | intern2 | | [Llama](https://github.com/facebookresearch/llama) | 7B/13B/33B/65B | - | | [Llama 2](https://huggingface.co/meta-llama) | 7B/13B/70B | llama2 | -| [Llama 3](https://huggingface.co/meta-llama) | 8B/70B | llama3 | +| [Llama 3/Llama 3.1](https://huggingface.co/meta-llama) | 8B/70B | llama3 | | [LLaVA-1.5](https://huggingface.co/llava-hf) | 7B/13B | vicuna | | [Mistral/Mixtral](https://huggingface.co/mistralai) | 7B/8x7B/8x22B | mistral | | [OLMo](https://huggingface.co/allenai) | 1B/7B | - | diff --git a/README_zh.md b/README_zh.md index c82d95d2..0758c622 100644 --- a/README_zh.md +++ b/README_zh.md @@ -164,7 +164,7 @@ https://github.com/hiyouga/LLaMA-Factory/assets/16256802/ec36a9dd-37f4-4f72-81bd | [InternLM2](https://huggingface.co/internlm) | 7B/20B | intern2 | | [Llama](https://github.com/facebookresearch/llama) | 7B/13B/33B/65B | - | | [Llama 2](https://huggingface.co/meta-llama) | 7B/13B/70B | llama2 | -| [Llama 3](https://huggingface.co/meta-llama) | 8B/70B | llama3 | +| [Llama 3/Llama 3.1](https://huggingface.co/meta-llama) | 8B/70B | llama3 | | [LLaVA-1.5](https://huggingface.co/llava-hf) | 7B/13B | vicuna | | [Mistral/Mixtral](https://huggingface.co/mistralai) | 7B/8x7B/8x22B | mistral | | [OLMo](https://huggingface.co/allenai) | 1B/7B | - | diff --git a/src/llamafactory/extras/constants.py b/src/llamafactory/extras/constants.py index ac082043..2190292c 100644 --- a/src/llamafactory/extras/constants.py +++ b/src/llamafactory/extras/constants.py @@ -739,6 +739,35 @@ register_model_group( ) +register_model_group( + models={ + "LLaMA3.1-8B": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-8B", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-8B", + }, + "LLaMA3.1-70B": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-70B", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-70B", + }, + "LLaMA3.1-405B": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-405B", + }, + "LLaMA3.1-8B-Chat": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-8B-Instruct", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-8B-Instruct", + }, + "LLaMA3.1-70B-Chat": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-70B-Instruct", + DownloadSource.MODELSCOPE: "LLM-Research/Meta-Llama-3.1-70B-Instruct", + }, + "LLaMA3.1-405B-Chat": { + DownloadSource.DEFAULT: "meta-llama/Meta-Llama-3.1-405B-Instruct", + }, + }, + template="llama3", +) + + register_model_group( models={ "LLaVA1.5-7B-Chat": {