mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 22:32:54 +08:00
try to past test
Former-commit-id: 7b4ba0efb658422fd29dca63bac1e9cee8e82af8
This commit is contained in:
parent
677d57b7c7
commit
16c7326bc5
@ -161,7 +161,7 @@ Compared to ChatGLM's [P-Tuning](https://github.com/THUDM/ChatGLM2-6B/tree/main/
|
|||||||
## Supported Models
|
## Supported Models
|
||||||
|
|
||||||
| Model | Model size | Template |
|
| Model | Model size | Template |
|
||||||
|-------------------------------------------------------------------| -------------------------------- |------------------|
|
| ----------------------------------------------------------------- | -------------------------------- | --------- |
|
||||||
| [Baichuan 2](https://huggingface.co/baichuan-inc) | 7B/13B | baichuan2 |
|
| [Baichuan 2](https://huggingface.co/baichuan-inc) | 7B/13B | baichuan2 |
|
||||||
| [BLOOM/BLOOMZ](https://huggingface.co/bigscience) | 560M/1.1B/1.7B/3B/7.1B/176B | - |
|
| [BLOOM/BLOOMZ](https://huggingface.co/bigscience) | 560M/1.1B/1.7B/3B/7.1B/176B | - |
|
||||||
| [ChatGLM3](https://huggingface.co/THUDM) | 6B | chatglm3 |
|
| [ChatGLM3](https://huggingface.co/THUDM) | 6B | chatglm3 |
|
||||||
|
@ -379,6 +379,7 @@ class LlavaNextVideoPlugin(BasePlugin):
|
|||||||
res.update(video_res)
|
res.update(video_res)
|
||||||
return res
|
return res
|
||||||
|
|
||||||
|
|
||||||
class PaliGemmaPlugin(BasePlugin):
|
class PaliGemmaPlugin(BasePlugin):
|
||||||
@override
|
@override
|
||||||
def process_messages(
|
def process_messages(
|
||||||
|
@ -120,9 +120,12 @@ def load_config(model_args: "ModelArguments") -> "PretrainedConfig":
|
|||||||
"""
|
"""
|
||||||
init_kwargs = _get_init_kwargs(model_args)
|
init_kwargs = _get_init_kwargs(model_args)
|
||||||
if "LLaVA-NeXT-Video" in model_args.model_name_or_path:
|
if "LLaVA-NeXT-Video" in model_args.model_name_or_path:
|
||||||
from transformers import PretrainedConfig, LlavaNextVideoConfig, CLIPVisionConfig, LlamaConfig
|
from transformers import CLIPVisionConfig, LlamaConfig, LlavaNextVideoConfig, PretrainedConfig
|
||||||
|
|
||||||
official_config = PretrainedConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs)
|
official_config = PretrainedConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs)
|
||||||
config = LlavaNextVideoConfig(CLIPVisionConfig(**official_config.vision_config), LlamaConfig(**official_config.text_config))
|
config = LlavaNextVideoConfig(
|
||||||
|
CLIPVisionConfig(**official_config.vision_config), LlamaConfig(**official_config.text_config)
|
||||||
|
)
|
||||||
setattr(config, "visual_inputs", True)
|
setattr(config, "visual_inputs", True)
|
||||||
return config
|
return config
|
||||||
return AutoConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs)
|
return AutoConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs)
|
||||||
@ -163,6 +166,7 @@ def load_model(
|
|||||||
load_class = AutoModelForCausalLM
|
load_class = AutoModelForCausalLM
|
||||||
if "llava_next_video" == getattr(config, "model_type"):
|
if "llava_next_video" == getattr(config, "model_type"):
|
||||||
from transformers import LlavaNextVideoForConditionalGeneration
|
from transformers import LlavaNextVideoForConditionalGeneration
|
||||||
|
|
||||||
load_class = LlavaNextVideoForConditionalGeneration
|
load_class = LlavaNextVideoForConditionalGeneration
|
||||||
|
|
||||||
if model_args.train_from_scratch:
|
if model_args.train_from_scratch:
|
||||||
|
@ -108,7 +108,13 @@ def configure_visual_model(config: "PretrainedConfig") -> None:
|
|||||||
Patches VLMs before loading them.
|
Patches VLMs before loading them.
|
||||||
"""
|
"""
|
||||||
model_type = getattr(config, "model_type", None)
|
model_type = getattr(config, "model_type", None)
|
||||||
if model_type in ["llava", "llava_next", "video_llava", "idefics2", "llava_next_video"]: # required for ds zero3 and valuehead models
|
if model_type in [
|
||||||
|
"llava",
|
||||||
|
"llava_next",
|
||||||
|
"video_llava",
|
||||||
|
"idefics2",
|
||||||
|
"llava_next_video",
|
||||||
|
]: # required for ds zero3 and valuehead models
|
||||||
setattr(config, "hidden_size", getattr(config.text_config, "hidden_size", None))
|
setattr(config, "hidden_size", getattr(config.text_config, "hidden_size", None))
|
||||||
|
|
||||||
if getattr(config, "is_yi_vl_derived_model", None):
|
if getattr(config, "is_yi_vl_derived_model", None):
|
||||||
|
Loading…
x
Reference in New Issue
Block a user