423A35C7
  • Joined on 2024-05-11
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-14 19:06:26 +08:00
aeda079014 [v1] model loader (#9613)
423A35C7 synced and deleted reference refs/tags/hiyouga/mo at 423A35C7/LLaMA-Factory from mirror 2025-12-14 19:06:26 +08:00
423A35C7 synced new reference hiyouga/mo to 423A35C7/LLaMA-Factory from mirror 2025-12-14 10:56:56 +08:00
423A35C7 synced commits to hiyouga/mo at 423A35C7/LLaMA-Factory from mirror 2025-12-14 10:56:56 +08:00
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-14 02:46:58 +08:00
fdd24276ed [feat] support new function call value (#9610)
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-13 02:16:57 +08:00
110d21713e [v1] add dp & mp mesh (#9611)
203069e11c [v1] add accelerator (#9607)
Compare 2 commits »
423A35C7 synced and deleted reference refs/tags/hiyouga/acc at 423A35C7/LLaMA-Factory from mirror 2025-12-13 02:16:57 +08:00
423A35C7 synced commits to hiyouga/acc at 423A35C7/LLaMA-Factory from mirror 2025-12-12 18:08:19 +08:00
3b06033357 add accelerator
423A35C7 synced new reference hiyouga/acc to 423A35C7/LLaMA-Factory from mirror 2025-12-12 10:04:22 +08:00
423A35C7 synced commits to hiyouga/acc at 423A35C7/LLaMA-Factory from mirror 2025-12-12 10:04:22 +08:00
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-10 17:14:17 +08:00
4fd94141a4 [model] Add Ministral3 (#9582)
22d6ac29d5 [model] Rename GLMV template (#9595)
Compare 2 commits »
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-10 00:54:21 +08:00
cff4483392 [config] Fix RoPE scaling patch for resuming from a scaled model (#9588)
5d56817e2b [misc] lint (#9593)
Compare 2 commits »
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-09 16:44:20 +08:00
1bbb461f76 [assets] update readme (#9587)
c1f5f8fff6 [model] support GLM4.6v (#9586)
Compare 2 commits »
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-08 08:04:17 +08:00
5744f1ea94 [v1] add models & accelerator (#9579)
739954910a [deps] Update for Transformers v5 (#9569)
Compare 2 commits »
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-06 23:24:20 +08:00
109162dc56 [fix] fix the issue when using fsdp2 with gradient checkpointing. (#9541)
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-06 07:04:18 +08:00
165f3f073a [examples] add fsdp config for mutiple nodes (#9575)
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-12-02 05:04:19 +08:00
efb13b7483 [V1] Refactor ascend MoE kernel patch logic & Support Qwen3-MoE (#9557)
423A35C7 synced commits to main at 423A35C7/LLaMA-Factory from mirror 2025-11-30 12:14:17 +08:00
e43a972b25 [test] add npu test yaml and add ascend a3 docker file (#9547)
22be45c78c [misc] fix omni thinker load (#9552)
Compare 2 commits »
423A35C7 synced new reference refs/tags/v0.7.9 to 423A35C7/pytorch3d from mirror 2025-11-29 03:54:18 +08:00
423A35C7 synced commits to refs/tags/v0.7.9 at 423A35C7/pytorch3d from mirror 2025-11-29 03:54:18 +08:00