Commit Graph

67 Commits

Author SHA1 Message Date
hiyouga
88c7fc1599 set dev version 2024-07-19 02:01:46 +08:00
hiyouga
bbd5a64423 release v0.8.3 2024-07-19 01:21:18 +08:00
hiyouga
d774b94f12 support batch_eval_metrics, fix #4826 2024-07-17 00:33:00 +08:00
hoshi-hiyouga
f84b007ebb Update packages.py 2024-07-07 15:48:29 +08:00
Lian Junhong
322663bf90 chore: Update vllm_engine.py to support vllm version >= 0.5.1 2024-07-07 15:08:12 +08:00
hiyouga
53b1002fb7 add codegeex4, internlm2.5 2024-07-06 16:16:47 +08:00
hiyouga
6fd6aa4530 fix packing for eager/sdpa attn 2024-07-04 01:52:43 +08:00
hoshi-hiyouga
87d9b2d005 Merge pull request #4224 from chuan298/main
Implement efficient packing without cross-contamination attention
2024-07-04 01:18:54 +08:00
hiyouga
cce7083024 update packing 2024-07-04 01:10:55 +08:00
hiyouga
8a6a7b9c8a update arg name 2024-07-03 23:23:24 +08:00
hiyouga
c47ab6c072 improve rlhf 2024-07-02 22:23:08 +08:00
hzhaoy
57b7c00430 add TeleChat-1B 2024-07-02 17:49:04 +08:00
hoshi-hiyouga
e8e6af2651 Merge branch 'main' into main 2024-07-01 21:01:09 +08:00
hiyouga
d74244d568 fix #4398 #4592 2024-06-30 21:28:51 +08:00
hiyouga
6f63050e1b add Gemma2 models 2024-06-28 01:26:50 +08:00
hiyouga
8baf3b22b0 refactor pissa, improve llamaboard 2024-06-28 01:04:24 +08:00
hiyouga
ad144c2265 support HQQ/EETQ #4113 2024-06-27 00:29:42 +08:00
hiyouga
e507e60638 update readme 2024-06-24 18:22:12 +08:00
ancv
770f75dc83 move configure_packing to llamafactory.model.patcher and fix constants 2024-06-21 00:45:06 +07:00
hiyouga
42e69a3c63 set dev version 2024-06-19 21:08:16 +08:00
hiyouga
71327ba85a release v0.8.2 2024-06-19 20:42:09 +08:00
hiyouga
a233fbc258 add deepseek coder v2 #4346 2024-06-18 22:53:54 +08:00
ancv
238f5c3d99 update packing with sdpa and eager attention mode 2024-06-16 02:25:47 +07:00
hiyouga
38b6b0f52e tiny fix 2024-06-16 01:06:41 +08:00
hiyouga
1b834f50be add tests 2024-06-15 19:51:20 +08:00
hiyouga
572d8bbfdd add minicpm #4227 2024-06-15 17:58:52 +08:00
hiyouga
d87108daa6 add license 2024-06-15 17:54:33 +08:00
hiyouga
2ed8270112 clean code 2024-06-13 01:58:16 +08:00
hiyouga
91e62a098f set dev version 2024-06-11 00:50:53 +08:00
hiyouga
2b6ebd6b51 release v0.8.1 2024-06-11 00:44:26 +08:00
hiyouga
972ec9c668 fix llamafactory-cli env 2024-06-08 07:15:45 +08:00
hiyouga
3ac11e77cc set dev version 2024-06-08 06:46:09 +08:00
hiyouga
5aa4ce4756 release v0.8.0 2024-06-08 05:20:54 +08:00
hiyouga
06e5d136a4 add resume args in webui 2024-06-08 00:22:16 +08:00
hiyouga
f9e818d79c fix #4120 2024-06-07 04:18:05 +08:00
hiyouga
8e95648850 add qwen2 models 2024-06-07 00:22:57 +08:00
hiyouga
451b6693c0 fix torch gc 2024-06-06 20:30:25 +08:00
hiyouga
cae4737907 lora modules: all by default 2024-06-06 03:53:28 +08:00
hiyouga
c23cc63d3d add codestral 22B 2024-06-06 03:42:50 +08:00
hiyouga
7daf8366db lint 2024-06-06 03:33:44 +08:00
hoshi-hiyouga
f2580ad403 Merge pull request #4066 from injet-zhou/main
add throughput entry to training log
2024-06-06 03:32:04 +08:00
hiyouga
dc4a00dd63 update train hparams 2024-06-06 01:49:20 +08:00
hiyouga
d4908d5708 add llamafactory-cli env 2024-06-06 01:28:14 +08:00
hiyouga
67fe822324 fix #4090 2024-06-06 00:50:32 +08:00
hiyouga
f48f5e646e support glm-4 2024-06-05 15:16:38 +08:00
faddddeout
b2f0459542 add throughput entry to log 2024-06-04 11:04:29 +00:00
hiyouga
876bc92865 bump versions
transformers 4.37.2->4.41.2
datasets 2.14.3->2.16.0
accelerate 0.27.2->0.30.1
peft 0.10.0->0.11.1
trl 0.8.1->0.8.6
2024-06-03 18:29:38 +08:00
hiyouga
8070871732 better llamaboard
* easily resume from checkpoint
* support full and freeze checkpoints
* faster ui
2024-05-29 23:55:38 +08:00
hiyouga
89ca832740 update readme 2024-05-29 18:39:11 +08:00
hzhaoy
0dd632fe9e add TeleChat-12B/TeleChat-12B-v2 models 2024-05-29 15:00:37 +08:00