Commit Graph

1918 Commits

Author SHA1 Message Date
hiyouga
f1d8d29bc3 add gemma test 2024-07-14 18:01:45 +08:00
hiyouga
173921419d fix test 2024-07-14 15:44:30 +08:00
hiyouga
88a20ba797 fix #4699
slow tokenizer for yi models
2024-07-14 15:34:22 +08:00
hiyouga
d3c01552e0 tiny fix 2024-07-14 10:56:45 +08:00
hiyouga
2f6af73da2 fix gemma2 attention 2024-07-13 23:33:45 +08:00
hiyouga
7b19e99ed7 update workflows 2024-07-13 22:31:15 +08:00
hoshi-hiyouga
5da54deb50 Merge pull request #4781 from hzhaoy/fix-dockerfile-cuda
Fix cuda Dockerfile
2024-07-13 22:25:32 +08:00
hiyouga
6b48308ef9 fix #4792 2024-07-13 22:07:58 +08:00
hoshi-hiyouga
32699a82a6 Merge pull request #4804 from codemayq/fix-examples
tiny fix of examples
2024-07-13 20:49:13 +08:00
hoshi-hiyouga
f618b80fa2 Update llava1_5.yaml 2024-07-13 20:30:06 +08:00
codingma
982a1cdd24 1. fix output_dir in llama3_lora_pretrain.yaml
2. add llava1_5.yaml for inference
2024-07-13 13:16:22 +08:00
hzhaoy
8bab99c582 tiny fix 2024-07-12 00:28:44 +08:00
hzhaoy
642c6d666f fix #4780 2024-07-12 00:25:48 +08:00
hzhaoy
a8bf1abf0f fix #4779 2024-07-12 00:15:15 +08:00
codemayq
67040f149c update wechat_npu.jpg 2024-07-11 20:03:39 +08:00
hoshi-hiyouga
555194e150 Merge pull request #4700 from marko1616/patch-1
Fix Windows command preview
2024-07-10 13:51:50 +08:00
hoshi-hiyouga
40c3b88b68 Merge pull request #4746 from yzoaim/fix
fix src/llamafactory/train/callbacks.py
2024-07-10 13:32:49 +08:00
hoshi-hiyouga
39cd89ce17 Update callbacks.py 2024-07-10 13:32:20 +08:00
-.-
cff89a2e89 fix src/llamafactory/train/callbacks.py 2024-07-10 12:05:51 +08:00
hiyouga
51942acee8 fix #4731 2024-07-10 11:32:36 +08:00
hiyouga
fb0c400116 fix ppo trainer 2024-07-10 11:05:45 +08:00
hiyouga
2f09520c0d fix #4742 2024-07-09 23:24:24 +08:00
hiyouga
86b1594823 Update wechat.jpg 2024-07-09 09:25:11 +08:00
hoshi-hiyouga
563a27dab7 Merge pull request #4706 from T-Atlas/main
chore: Update vllm_engine.py to support vllm version >= 0.5.1
2024-07-07 15:50:38 +08:00
hoshi-hiyouga
f84b007ebb Update packages.py 2024-07-07 15:48:29 +08:00
Lian Junhong
322663bf90 chore: Update vllm_engine.py to support vllm version >= 0.5.1 2024-07-07 15:08:12 +08:00
hiyouga
a15782cb9f fix #4705 2024-07-07 13:10:06 +08:00
marko1616
e0562521bb Update utils.py
In windows mutiline command should like
command --arg1 xxx `
--arg2 xxx `
2024-07-06 20:40:13 +08:00
hiyouga
53b1002fb7 add codegeex4, internlm2.5 2024-07-06 16:16:47 +08:00
hiyouga
c9bb0757ec update pissa example 2024-07-06 15:47:32 +08:00
hiyouga
9f33f1edf5 fix processors 2024-07-05 08:33:22 +08:00
hiyouga
e43809bced fix #4683 2024-07-05 00:58:05 +08:00
hiyouga
ed232311e8 fix #4674 2024-07-05 00:41:03 +08:00
hiyouga
226a9e563f Merge branch 'main' of https://github.com/hiyouga/LLaMA-Factory 2024-07-04 14:23:37 +08:00
hiyouga
1e27e8c776 fix #4677 2024-07-04 14:22:07 +08:00
hoshi-hiyouga
07d96d497c Merge pull request #4673 from hzhaoy/main
tiny fix
2024-07-04 10:40:41 +08:00
hzhaoy
738df47748 tiny fix 2024-07-04 10:20:28 +08:00
hiyouga
636bb9c1e6 update tests 2024-07-04 04:00:12 +08:00
hiyouga
0c699de39d tiny fix 2024-07-04 03:47:05 +08:00
hiyouga
44747cebd2 tiny fix 2024-07-04 03:02:23 +08:00
hiyouga
b5d101e1bf fix data map for packing 2024-07-04 03:01:31 +08:00
hiyouga
b03e4a74ba update wechat 2024-07-04 01:55:05 +08:00
hiyouga
6fd6aa4530 fix packing for eager/sdpa attn 2024-07-04 01:52:43 +08:00
hoshi-hiyouga
87d9b2d005 Merge pull request #4224 from chuan298/main
Implement efficient packing without cross-contamination attention
2024-07-04 01:18:54 +08:00
hiyouga
cce7083024 update packing 2024-07-04 01:10:55 +08:00
hoshi-hiyouga
a36e8f2dd5 Update packing.py 2024-07-03 23:36:01 +08:00
hiyouga
c346f79f99 update func name 2024-07-03 23:29:33 +08:00
hiyouga
8a6a7b9c8a update arg name 2024-07-03 23:23:24 +08:00
hiyouga
575a02a23d update hparams 2024-07-03 23:18:58 +08:00
hiyouga
7f770f6895 update ui 2024-07-03 23:13:49 +08:00