Logo
Explore Help
Register Sign In
423A35C7/LLaMA-Factory
1
0
Fork 0
You've already forked LLaMA-Factory
mirror of https://github.com/hiyouga/LLaMA-Factory.git synced 2025-11-08 14:24:47 +08:00
Code Issues Packages Projects Releases Wiki Activity
LLaMA-Factory/src/llamafactory/model/model_utils
History
Kingsley 8efa5c9694
[model] support GLM4.5V (#8876)
2025-08-11 21:45:14 +08:00
..
__init__.py
rename files
2024-06-07 00:09:06 +08:00
attention.py
[model] switch to gptqmodel (#8108)
2025-05-19 22:25:40 +08:00
checkpointing.py
[model] fix vit gradient checkpointing (#7830)
2025-04-23 22:48:48 +08:00
embedding.py
[train] fix adjusting logits size after adding special tokens (#8823)
2025-08-05 20:35:07 +08:00
kv_cache.py
[model] fix kv cache (#7564)
2025-04-01 23:07:46 +08:00
liger_kernel.py
[version] release v0.9.3 (#8386)
2025-06-16 19:21:32 +08:00
longlora.py
[trainer] fix pt loss (#7748)
2025-04-17 03:15:35 +08:00
misc.py
[misc] upgrade format to py39 (#7256)
2025-03-12 00:08:41 +08:00
mod.py
[misc] update license year & fix llama pro (#6814)
2025-02-05 01:53:33 +08:00
moe.py
[model] support GLM4.5V (#8876)
2025-08-11 21:45:14 +08:00
packing.py
[trainer] fix pt loss (#7748)
2025-04-17 03:15:35 +08:00
quantization.py
[data] fix template (#8827)
2025-08-06 06:58:09 +08:00
rope.py
[model] switch to gptqmodel (#8108)
2025-05-19 22:25:40 +08:00
unsloth.py
Revert "[model] add lora dropout to unsloth" - requested feature already exists (#8554)
2025-07-05 11:25:31 +08:00
valuehead.py
Merge commit from fork
2025-06-26 13:55:42 +08:00
visual.py
[model] support GLM4.5V (#8876)
2025-08-11 21:45:14 +08:00
Powered by Gitea Version: 23.8.0 Page: 76ms Template: 2ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API