Kingsley
|
13170577b2
|
[feat] support megatron-LM training by mcore_adapter (#9237)
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
Co-authored-by: Yaowei Zheng <hiyouga@buaa.edu.cn>
|
2025-10-26 16:21:30 +08:00 |
|
Yaowei Zheng
|
47a7dc1698
|
[deps] upgrade vllm (#9293)
|
2025-10-16 23:20:26 +08:00 |
|
Ben Feuer
|
1c44b60e3e
|
[feat] fp8 training (#8960)
Co-authored-by: Benjamin Feuer <penfever@gmail.com>
Co-authored-by: Yaowei Zheng <hiyouga@buaa.edu.cn>
|
2025-10-01 14:32:53 +08:00 |
|
Yaowei Zheng
|
6ffebe5ff7
|
[data] fix qwen omni plugin (#9204)
Co-authored-by: kingsley <kingsleydodonow@gmail.com>
|
2025-09-28 01:02:29 +08:00 |
|
Yaowei Zheng
|
2c31279316
|
[assets] update wechat (#8962)
|
2025-08-19 02:55:09 +08:00 |
|
Zeju Qiu
|
003a2acb1a
|
[feature] adding orthogononal finetuning (OFT) to llama factory (#8623)
Co-authored-by: Zeju <zqiu@g003.internal.cluster.is.localnet>
Co-authored-by: Zeju <zqiu@login2.is.localnet>
Co-authored-by: Yaowei Zheng <hiyouga@buaa.edu.cn>
|
2025-08-18 18:22:47 +08:00 |
|
Yaowei Zheng
|
a416ab48d8
|
[deps] upgrade vllm to 0.10.0 (#8787)
|
2025-07-30 22:26:38 +08:00 |
|
Yaowei Zheng
|
4b0ec83928
|
[deps] bump transformers to 4.49.0 (#8564)
|
2025-07-07 20:31:50 +08:00 |
|
Ze-Yi LIN
|
16f13d304b
|
[tracking] fix swanlab hparams (#8532)
Co-authored-by: Yaowei Zheng <hiyouga@buaa.edu.cn>
|
2025-07-02 22:08:44 +08:00 |
|
Kingsley
|
bede213da7
|
[assets] update readme (#8519)
|
2025-07-02 15:38:38 +08:00 |
|
Injae Ryou
|
a5a93597b1
|
[parser] update config loading to use OmegaConf #7793 (#8505)
|
2025-07-01 21:05:13 +08:00 |
|
Yaowei Zheng
|
c6c764388c
|
[assets] update readme (#8396)
|
2025-06-17 16:15:20 +08:00 |
|
Yaowei Zheng
|
3a3bae1cfe
|
[data] fix qwen2vl pos ids (#8387)
|
2025-06-17 00:48:54 +08:00 |
|
hoshi-hiyouga
|
45030ff803
|
[model] switch to gptqmodel (#8108)
|
2025-05-19 22:25:40 +08:00 |
|
hoshi-hiyouga
|
73198a6645
|
[misc] fix uv (#7913)
|
2025-04-30 07:45:03 +08:00 |
|
hoshi-hiyouga
|
b07628dea5
|
[example] add bash usage (#7794)
|
2025-04-22 00:25:51 +08:00 |
|
Juanxi Tian
|
12ada72ed4
|
[trainer] Add Muon Optimizer (#7749)
Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
|
2025-04-21 23:38:37 +08:00 |
|
hoshi-hiyouga
|
416853dd25
|
[parser] support omegaconf (#7793)
|
2025-04-21 23:30:30 +08:00 |
|
hoshi-hiyouga
|
d222f63cb7
|
[infer] set env for vllm ascend (#7745)
|
2025-04-17 01:08:55 +08:00 |
|
hoshi-hiyouga
|
3df021d4d7
|
[deps] upgrade vllm (#7728)
|
2025-04-15 14:57:40 +08:00 |
|
hoshi-hiyouga
|
7c61b35106
|
[misc] upgrade cli (#7714)
|
2025-04-14 15:41:22 +08:00 |
|
hoshi-hiyouga
|
c3c0efbaa0
|
[misc] fix packing and eval plot (#7623)
|
2025-04-07 18:20:57 +08:00 |
|
hoshi-hiyouga
|
2bfcad2394
|
[model] fix kv cache (#7564)
|
2025-04-01 23:07:46 +08:00 |
|
hoshi-hiyouga
|
919415dba9
|
[deps] upgrade vllm to 0.8 (#7436)
|
2025-03-23 14:32:22 +08:00 |
|
Qiaolin Yu
|
a44a53ebec
|
[inference] support sglang backend (#7278)
* Mimic SGLang offline Engine
* Add more tests and args
* Pass all current tests
* Clean Code
* fix sample_params
* clean code
* Fix Stream Chat
* change sglang from engine mode to server mode
* fix
* Fix Review Issues
* Use SGLang Built-In Utilities
* Fix test SGLang
* Some Doc Issue
* fix sglang engine
* add readme
---------
Co-authored-by: Jin Pan <jpan236@wisc.edu>
Co-authored-by: hiyouga <hiyouga@buaa.edu.cn>
|
2025-03-15 04:37:58 +08:00 |
|
hoshi-hiyouga
|
650a9a9057
|
[misc] update format (#7277)
|
2025-03-13 02:53:08 +08:00 |
|
hoshi-hiyouga
|
264538cb26
|
[misc] upgrade format to py39 (#7256)
|
2025-03-12 00:08:41 +08:00 |
|
hoshi-hiyouga
|
cf58a6d860
|
[config] fix export max len (#7230)
Former-commit-id: 211c0b3e8f3340acd2fae1762d9152a09f19ba34
|
2025-03-10 16:46:08 +08:00 |
|
hoshi-hiyouga
|
a255c3a476
|
[misc] fix cli (#7204)
Former-commit-id: 999f57133ca163c7108d2d5ee8194eca9b2109b4
|
2025-03-07 15:01:18 +08:00 |
|
hoshi-hiyouga
|
d2f845d70d
|
[deps] upgrade vllm (#7183)
Former-commit-id: 37678a3d64668c3b4a4bfefc054e3b9b40427c1a
|
2025-03-06 15:25:08 +08:00 |
|
hoshi-hiyouga
|
3a3f4072e5
|
[misc] fix grad ckpt func (#6916)
Former-commit-id: 35e069a52b3d7cfd9b0107574b09265eb2290f0b
|
2025-02-13 00:17:18 +08:00 |
|
hoshi-hiyouga
|
2581cc844b
|
[data] feat: auto template (#6905)
* support auto template
* add unittest
Former-commit-id: 0c6c9150db6414a5a05527ea486dce6633dff4b3
|
2025-02-12 00:22:53 +08:00 |
|
hoshi-hiyouga
|
4d1791e905
|
[deps] upgrade vllm (#6857)
Former-commit-id: 4bd50f65a3d62528768561019fda2723d045c7fd
|
2025-02-08 15:02:28 +08:00 |
|
hoshi-hiyouga
|
fec641ec82
|
[misc] allow extra args (#6831)
Former-commit-id: 0fd3a5295cb4e08a4e57e860e82103364c28fba8
|
2025-02-06 12:38:08 +08:00 |
|
hoshi-hiyouga
|
7638f1070e
|
[optim] clean apollo (#6645)
* clean apollo code
* update readme
Former-commit-id: 38b8ec4a99189483124b54df9d6bc6b0d318855a
|
2025-01-15 01:42:50 +08:00 |
|
zhuHQ
|
c2120432db
|
[optim] add support to APOLLO (#6617)
Former-commit-id: 5a252e5a458457adbd19da3b68a3897ad2962824
|
2025-01-15 00:24:56 +08:00 |
|
hoshi-hiyouga
|
28d145a066
|
pin vllm version to 0.6.5 (#6629)
Former-commit-id: 26097ca0adf25ebb7d9e8eec2d2cef673c6cfe88
|
2025-01-14 02:44:02 +08:00 |
|
hoshi-hiyouga
|
4e25d037c8
|
Merge pull request #6564 from stephen-nju/fix_ray
Fix ray
Former-commit-id: d4566839369726023f1b6e8f4b2332bda0c715cc
|
2025-01-08 18:14:18 +08:00 |
|
zhubin
|
b6b53b61f7
|
fix get ray args when args not a dict
Former-commit-id: 5e5398cd5b117b2378107172d3f91cfb0321e842
|
2025-01-08 10:06:02 +00:00 |
|
hiyouga
|
647c51a772
|
imporve log
Former-commit-id: a6abf375975ffea3d51e1b944c9855b5f62ffac8
|
2025-01-08 09:56:10 +00:00 |
|
hiyouga
|
944a2aec4d
|
refactor ray integration, support save ckpt
Former-commit-id: 2f50b27e608b2092bfceab6c6e84e6631e973ee2
|
2025-01-07 09:39:10 +00:00 |
|
Eric Tang
|
4f31ad997c
|
run style check
Former-commit-id: 5ec33baf5f95df9fa2afe5523c825d3eda8a076b
|
2025-01-07 08:55:44 +00:00 |
|
Kourosh Hakhamaneshi
|
8683582300
|
drafting ray integration
Signed-off-by: Kourosh Hakhamaneshi <kourosh@anyscale.com>
Former-commit-id: 19c12ddae9350f6e25a270fe3372f5b9094cf960
|
2025-01-07 08:55:44 +00:00 |
|
hiyouga
|
f8f05a883b
|
fix #6482
Former-commit-id: 8577f52b4152efe6cc7a8b5f6d37b4f9ba6684e7
|
2024-12-30 06:03:07 +00:00 |
|
hiyouga
|
c1768cfb14
|
support batch infer in vllm
Former-commit-id: 3ef5ed3b9a44eed2f7e3ff221dfc343d0a97c0b5
|
2024-12-04 13:50:00 +00:00 |
|
hiyouga
|
1e6f96508a
|
add vllm config
Former-commit-id: 95365f0ce4f362bde7de8b679b54b548d7055bfb
|
2024-11-10 21:28:18 +08:00 |
|
hiyouga
|
093eda2ad6
|
support rank0 logger
Former-commit-id: 84528eabe560091bfd866b6a0ca864085af7529b
|
2024-11-02 18:31:04 +08:00 |
|
hiyouga
|
248d5daaff
|
use pre-commit
Former-commit-id: 7cfede95df22a9ff236788f04159b6b16b8d04bb
|
2024-10-29 09:07:46 +00:00 |
|
hiyouga
|
c7efc7f2ed
|
tiny fix
Former-commit-id: 1fe424323b212094856f423351dc2a15774d39c3
|
2024-10-11 23:51:54 +08:00 |
|
Johnny
|
9d27aaa38f
|
Update parser.py
Former-commit-id: 60b13c86f4feaffbb43f5a23a28376fe416ed118
|
2024-10-11 12:29:33 +02:00 |
|