mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2026-03-01 17:26:03 +08:00
[ci] disable pip cache for ci (#9654)
This commit is contained in:
@@ -82,6 +82,11 @@ def add_z3_leaf_module(model: "PreTrainedModel") -> None:
|
||||
|
||||
_set_z3_leaf_modules(model, [Glm4vMoeTextMoE])
|
||||
|
||||
if model_type == "gpt_oss":
|
||||
from transformers.models.gpt_oss.modeling_gpt_oss import GptOssMLP
|
||||
|
||||
_set_z3_leaf_modules(model, [GptOssMLP])
|
||||
|
||||
if model_type == "jamba":
|
||||
from transformers.models.jamba.modeling_jamba import JambaSparseMoeBlock
|
||||
|
||||
@@ -129,13 +134,9 @@ def add_z3_leaf_module(model: "PreTrainedModel") -> None:
|
||||
|
||||
if model_type in ("qwen3_omni_moe", "qwen3_omni_moe_thinker"):
|
||||
from transformers.models.qwen3_omni_moe.modeling_qwen3_omni_moe import Qwen3OmniMoeThinkerTextSparseMoeBlock
|
||||
|
||||
|
||||
_set_z3_leaf_modules(model, [Qwen3OmniMoeThinkerTextSparseMoeBlock])
|
||||
|
||||
if model_type == "gpt_oss":
|
||||
from transformers.models.gpt_oss.modeling_gpt_oss import GptOssMLP
|
||||
|
||||
_set_z3_leaf_modules(model, [GptOssMLP])
|
||||
|
||||
|
||||
def configure_moe(config: "PretrainedConfig", model_args: "ModelArguments", is_trainable: bool) -> None:
|
||||
if not is_trainable or not model_args.moe_aux_loss_coef:
|
||||
|
||||
Reference in New Issue
Block a user