2787 Commits

Author SHA1 Message Date
Juanxi Tian
d128382d3c
[trainer] Add Muon Optimizer (#7749)
Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-21 23:38:37 +08:00
hoshi-hiyouga
278df4308d
[parser] support omegaconf (#7793) 2025-04-21 23:30:30 +08:00
Changrui Chen
81768df04c
[data] Fix wrong position ids with packed attention masks (#7754)
Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-21 23:19:36 +08:00
flashJd
1302ca39f6
[misc] fix new tokens adding (#7253)
Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-21 23:19:02 +08:00
ddddng
b8cddbc7d7
[model] fix gemma3 export (#7786)
Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-21 23:07:11 +08:00
Sachin Beldona
ec7257e70f
[misc] fix bug in constant (#7765)
Co-authored-by: Sachin Beldona <sbeldona@cs.cmu.edu>
2025-04-21 23:06:31 +08:00
hoshi-hiyouga
a4455e3021
[assets] update wechat (#7792) 2025-04-21 21:29:42 +08:00
hoshi-hiyouga
610f164c69
[trainer] fix pt loss (#7748)
* fix pt loss

* robust

* fix

* test
2025-04-17 03:15:35 +08:00
hoshi-hiyouga
0a0cfeb782
[breaking] bump transformers to 4.45.0 & improve ci (#7746)
* update ci

* fix

* fix

* fix

* fix

* fix
2025-04-17 02:36:48 +08:00
hoshi-hiyouga
4831552856
[infer] set env for vllm ascend (#7745) 2025-04-17 01:08:55 +08:00
Kingsley
125513fa5c
[model] support intern-VL 2.5-3 series (#7258)
* add internvl and rebase

* fix for internvl2&3

* remove lines

* fix video_inputs & lint

* nit

* add constants

* remove lines

* fix

* fix error

* pass ci

* pass ci

* skip internvl & nit
2025-04-17 00:31:30 +08:00
ENg-122
8543400584
[misc] improve entrypoint (#7345)
* 纯粹优化下入口代码,因为看到if else太多了

* Update cli.py

---------

Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-16 21:48:23 +08:00
leo-pony
e1fdd6e2f8
[infer] support vllm-ascend (#7739) 2025-04-16 20:06:47 +08:00
codingma
d07983dceb
[assets] wechat (#7740) 2025-04-16 18:02:01 +08:00
hoshi-hiyouga
9b94211045
[api] fix chat messages (#7732) 2025-04-15 16:39:08 +08:00
hoshi-hiyouga
0fe5631f9b
[deps] upgrade vllm (#7728) 2025-04-15 14:57:40 +08:00
Joe Schoonover
b5d667cebf
[docker] patch docker-rocm (#7725)
* Update Dockerfile

* Fix typo

* Fix syntax for /bin/sh conditional

* Add build args to docker-compose

* Change shell to /bin/bash

This is required for "==" syntax in conditional string comparison
2025-04-15 13:36:39 +08:00
hoshi-hiyouga
ac8c6fdd3a
[assets] update model readme (#7724) 2025-04-15 00:41:09 +08:00
Kingsley
df8752e8ee
[model] Support Kimi_VL thinking/instruct (#7719)
* add kimi_vl

* patch config

* check version

* Update mm_plugin.py

* Update mm_plugin.py

---------

Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-15 00:21:58 +08:00
hoshi-hiyouga
3a13d2cdb1
[misc] fix env vars (#7715) 2025-04-14 16:04:04 +08:00
hoshi-hiyouga
3ef36d0057
[misc] upgrade cli (#7714) 2025-04-14 15:41:22 +08:00
hoshi-hiyouga
1fd4d14fbb
[deps] upgrade transformers (#7704) 2025-04-13 18:11:34 +08:00
Yuxuan Zhang
481ecbf9c5
[model] add GLM-4-0414 (#7695)
* Update README_zh.md

* update
2025-04-13 17:10:45 +08:00
hoshi-hiyouga
60a84f664b
[deps] fix uv conflicts (#7686)
* fix #7678

* Update setup.py

* Update tests.yml

* Update publish.yml

* Update Makefile
2025-04-11 18:02:24 +08:00
hoshi-hiyouga
11bcafd06a
[assets] update wechat (#7674) 2025-04-10 20:10:46 +08:00
Eric Tang
6c53471de2
[data] support for specifying a dataset in cloud storage (#7567)
* add support for loading datasets from s3/gcs

* add comments to readme

* run linter and address comments

* add option to pass in kwargs to ray init (i.e. runtime env)

* address comment

* revert mixed up changes
2025-04-10 11:31:35 +08:00
Eric Tang
39c1e29ed7
[ray] allow for specifying ray.init kwargs (i.e. runtime_env) (#7647)
* ray init kwargs

* Update trainer_utils.py

* fix ray args

---------

Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-10 11:31:05 +08:00
Dain Kim
ee840b4e01
[bugfix] enable_gemma_liger_kernel (#7660)
- The `enable_liger_kernel` function for the Gemma model series was not executed due to the existing `if` statement in the code.
- Changed the line to an `elif` statement so that the `apply_liger_kernel` function is executed properly.

resolved: #7628
2025-04-10 11:27:30 +08:00
jilongW
3bdc7e1e6c
[misc] fix cuda warn on intel GPU (#7655) 2025-04-09 21:37:54 +08:00
hoshi-hiyouga
34fdabe005
[data] add coig-p dataset (#7657) 2025-04-09 21:18:25 +08:00
hoshi-hiyouga
24cb890432
[assets] update readme (#7654) 2025-04-09 18:27:38 +08:00
hoshi-hiyouga
39876b85fc
[assets] update readme (#7644) 2025-04-09 01:06:06 +08:00
Kingsley
7d8bee96fc
[data] Fix bugs of use_audio_in_video in Qwen2.5 Omni (#7638)
* cache _mm_inputs

* nit

* support for use_audio_in_video

* remove cache

* fix data

* Update mllm_video_audio_demo.json
2025-04-08 18:40:10 +08:00
Shawn Tao
8f5f4cc559
[trainer] fix key error (#7635) 2025-04-08 18:39:50 +08:00
Adarsh Shirawalmath
8ee26642f3
[sglang] support transformers 4.51.0 (#7639) 2025-04-08 18:39:23 +08:00
hoshi-hiyouga
5817cda37e
[misc] fix packing and eval plot (#7623) 2025-04-07 18:20:57 +08:00
hoshi-hiyouga
7e0cdb1a76
[assets] update readme (#7612) 2025-04-06 13:58:49 +08:00
hoshi-hiyouga
6c200fd218
[model] add llama4 (#7611) 2025-04-06 13:42:31 +08:00
hoshi-hiyouga
61b24c3827
[assets] update wechat (#7594) 2025-04-03 17:45:26 +08:00
Kingsley
32cb086be1
[data] fix qwen2.5 omni plugin (#7578)
* specific entry

* Update mm_plugin.py

* fix fps cal

---------

Co-authored-by: hoshi-hiyouga <hiyouga@buaa.edu.cn>
2025-04-02 23:58:39 +08:00
Kingsley
80f8d037d0
[data] fix qwen2.5 omni plugin (#7573)
* align key with qwen2vl

* nit && change scripts
2025-04-02 21:28:52 +08:00
gechengze
11997593be
[trainer] fix batch processing in PPO trainer (#7576) 2025-04-02 21:17:48 +08:00
hoshi-hiyouga
903db09822
[infer] vllm video/audio inference (#7566) 2025-04-02 02:27:04 +08:00
hoshi-hiyouga
aaf2e6ba2a
[model] fix kv cache (#7564) 2025-04-01 23:07:46 +08:00
Yu Shi Jie
9deece1d50
[model] fix use_cache patching for gemma3 multimodal (#7500) 2025-04-01 16:06:48 +08:00
Ritesh Goru
f06a74ad4e
[data] specify position_ids in PackedSupervisedDatasetProcessor for neat_packing (#7318)
* use position_ids for neat_packing with fa2

* revert fa2 changes
2025-04-01 16:03:13 +08:00
taoharry
6faa6fb53d
[webui] fix launch with proxy (#7332) 2025-04-01 15:52:56 +08:00
Billy Cao
5d1cc863a4
[data] shard the dataset to allow multiprocessing when streaming is enabled (#7530)
* Shard the dataset when streaming to allow multiprocessing

* Allow user to not set dataset_shards to ensure backward compatibility
2025-04-01 15:36:23 +08:00
Hao
6d6e0f44fc
[trainer] new kto mismatch pair creation strategy (#7509) 2025-04-01 15:21:53 +08:00
hoshi-hiyouga
2d421c57bf
[data] fix qwen2.5 omni collator (#7553) 2025-04-01 00:15:12 +08:00