mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-11-05 02:12:14 +08:00
[model] add qwen3vl 2b & 32b (#9343)
This commit is contained in:
parent
2a822178de
commit
9c0d033a15
@ -5,7 +5,7 @@
|
|||||||
[](https://github.com/hiyouga/LLaMA-Factory/graphs/contributors)
|
[](https://github.com/hiyouga/LLaMA-Factory/graphs/contributors)
|
||||||
[](https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml)
|
[](https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml)
|
||||||
[](https://pypi.org/project/llamafactory/)
|
[](https://pypi.org/project/llamafactory/)
|
||||||
[](https://scholar.google.com/scholar?cites=12620864006390196564)
|
[](https://scholar.google.com/scholar?cites=12620864006390196564)
|
||||||
[](https://hub.docker.com/r/hiyouga/llamafactory/tags)
|
[](https://hub.docker.com/r/hiyouga/llamafactory/tags)
|
||||||
|
|
||||||
[](https://twitter.com/llamafactory_ai)
|
[](https://twitter.com/llamafactory_ai)
|
||||||
@ -317,7 +317,7 @@ Choose your path:
|
|||||||
| [Qwen2.5-Omni](https://huggingface.co/Qwen) | 3B/7B | qwen2_omni |
|
| [Qwen2.5-Omni](https://huggingface.co/Qwen) | 3B/7B | qwen2_omni |
|
||||||
| [Qwen3-Omni](https://huggingface.co/Qwen) | 30B | qwen3_omni |
|
| [Qwen3-Omni](https://huggingface.co/Qwen) | 30B | qwen3_omni |
|
||||||
| [Qwen2-VL/Qwen2.5-VL/QVQ](https://huggingface.co/Qwen) | 2B/3B/7B/32B/72B | qwen2_vl |
|
| [Qwen2-VL/Qwen2.5-VL/QVQ](https://huggingface.co/Qwen) | 2B/3B/7B/32B/72B | qwen2_vl |
|
||||||
| [Qwen3-VL](https://huggingface.co/Qwen) | 235B | qwen3_vl |
|
| [Qwen3-VL](https://huggingface.co/Qwen) | 2B/4B/8B/30B/32B/235B | qwen3_vl |
|
||||||
| [Seed (OSS/Coder)](https://huggingface.co/ByteDance-Seed) | 8B/36B | seed_oss/seed_coder |
|
| [Seed (OSS/Coder)](https://huggingface.co/ByteDance-Seed) | 8B/36B | seed_oss/seed_coder |
|
||||||
| [Skywork o1](https://huggingface.co/Skywork) | 8B | skywork_o1 |
|
| [Skywork o1](https://huggingface.co/Skywork) | 8B | skywork_o1 |
|
||||||
| [StarCoder 2](https://huggingface.co/bigcode) | 3B/7B/15B | - |
|
| [StarCoder 2](https://huggingface.co/bigcode) | 3B/7B/15B | - |
|
||||||
|
|||||||
@ -5,7 +5,7 @@
|
|||||||
[](https://github.com/hiyouga/LLaMA-Factory/graphs/contributors)
|
[](https://github.com/hiyouga/LLaMA-Factory/graphs/contributors)
|
||||||
[](https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml)
|
[](https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml)
|
||||||
[](https://pypi.org/project/llamafactory/)
|
[](https://pypi.org/project/llamafactory/)
|
||||||
[](https://scholar.google.com/scholar?cites=12620864006390196564)
|
[](https://scholar.google.com/scholar?cites=12620864006390196564)
|
||||||
[](https://hub.docker.com/r/hiyouga/llamafactory/tags)
|
[](https://hub.docker.com/r/hiyouga/llamafactory/tags)
|
||||||
|
|
||||||
[](https://twitter.com/llamafactory_ai)
|
[](https://twitter.com/llamafactory_ai)
|
||||||
@ -319,7 +319,7 @@ https://github.com/user-attachments/assets/43b700c6-a178-41db-b1f8-8190a5d3fcfc
|
|||||||
| [Qwen2.5-Omni](https://huggingface.co/Qwen) | 3B/7B | qwen2_omni |
|
| [Qwen2.5-Omni](https://huggingface.co/Qwen) | 3B/7B | qwen2_omni |
|
||||||
| [Qwen3-Omni](https://huggingface.co/Qwen) | 30B | qwen3_omni |
|
| [Qwen3-Omni](https://huggingface.co/Qwen) | 30B | qwen3_omni |
|
||||||
| [Qwen2-VL/Qwen2.5-VL/QVQ](https://huggingface.co/Qwen) | 2B/3B/7B/32B/72B | qwen2_vl |
|
| [Qwen2-VL/Qwen2.5-VL/QVQ](https://huggingface.co/Qwen) | 2B/3B/7B/32B/72B | qwen2_vl |
|
||||||
| [Qwen3-VL](https://huggingface.co/Qwen) | 235B | qwen3_vl |
|
| [Qwen3-VL](https://huggingface.co/Qwen) | 2B/4B/8B/30B/32B/235B | qwen3_vl |
|
||||||
| [Seed (OSS/Coder)](https://huggingface.co/ByteDance-Seed) | 8B/36B | seed_oss/seed_coder |
|
| [Seed (OSS/Coder)](https://huggingface.co/ByteDance-Seed) | 8B/36B | seed_oss/seed_coder |
|
||||||
| [Skywork o1](https://huggingface.co/Skywork) | 8B | skywork_o1 |
|
| [Skywork o1](https://huggingface.co/Skywork) | 8B | skywork_o1 |
|
||||||
| [StarCoder 2](https://huggingface.co/bigcode) | 3B/7B/15B | - |
|
| [StarCoder 2](https://huggingface.co/bigcode) | 3B/7B/15B | - |
|
||||||
|
|||||||
@ -3197,6 +3197,10 @@ register_model_group(
|
|||||||
|
|
||||||
register_model_group(
|
register_model_group(
|
||||||
models={
|
models={
|
||||||
|
"Qwen3-VL-2B-Instruct": {
|
||||||
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-2B-Instruct",
|
||||||
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-2B-Instruct",
|
||||||
|
},
|
||||||
"Qwen3-VL-4B-Instruct": {
|
"Qwen3-VL-4B-Instruct": {
|
||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-4B-Instruct",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-4B-Instruct",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-4B-Instruct",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-4B-Instruct",
|
||||||
@ -3205,6 +3209,10 @@ register_model_group(
|
|||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-8B-Instruct",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-8B-Instruct",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-8B-Instruct",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-8B-Instruct",
|
||||||
},
|
},
|
||||||
|
"Qwen3-VL-32B-Instruct": {
|
||||||
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-32B-Instruct",
|
||||||
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-32B-Instruct",
|
||||||
|
},
|
||||||
"Qwen3-VL-30B-A3B-Instruct": {
|
"Qwen3-VL-30B-A3B-Instruct": {
|
||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-30B-A3B-Instruct",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-30B-A3B-Instruct",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-30B-A3B-Instruct",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-30B-A3B-Instruct",
|
||||||
@ -3221,6 +3229,10 @@ register_model_group(
|
|||||||
|
|
||||||
register_model_group(
|
register_model_group(
|
||||||
models={
|
models={
|
||||||
|
"Qwen3-VL-2B-Thinking": {
|
||||||
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-2B-Thinking",
|
||||||
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-2B-Thinking",
|
||||||
|
},
|
||||||
"Qwen3-VL-4B-Thinking": {
|
"Qwen3-VL-4B-Thinking": {
|
||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-4B-Thinking",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-4B-Thinking",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-4B-Thinking",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-4B-Thinking",
|
||||||
@ -3229,6 +3241,10 @@ register_model_group(
|
|||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-8B-Thinking",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-8B-Thinking",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-8B-Thinking",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-8B-Thinking",
|
||||||
},
|
},
|
||||||
|
"Qwen3-VL-32B-Thinking": {
|
||||||
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-32B-Thinking",
|
||||||
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-32B-Thinking",
|
||||||
|
},
|
||||||
"Qwen3-VL-30B-A3B-Thinking": {
|
"Qwen3-VL-30B-A3B-Thinking": {
|
||||||
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-30B-A3B-Thinking",
|
DownloadSource.DEFAULT: "Qwen/Qwen3-VL-30B-A3B-Thinking",
|
||||||
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-30B-A3B-Thinking",
|
DownloadSource.MODELSCOPE: "Qwen/Qwen3-VL-30B-A3B-Thinking",
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user