mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-14 23:58:11 +08:00
Update README.md
Former-commit-id: e6fcc1831dadd2ec2c0acb14697a35f6471139ab
This commit is contained in:
parent
a6f800b741
commit
3d34d44497
@ -12,6 +12,10 @@
|
||||
|
||||
\[ English | [中文](README_zh.md) \]
|
||||
|
||||
Launch an all-in-one Web UI via `python src/train_web.py`.
|
||||
|
||||
https://github.com/hiyouga/LLaMA-Factory/assets/16256802/6ba60acc-e2e2-4bec-b846-2d88920d5ba1
|
||||
|
||||
## Changelog
|
||||
|
||||
[23/09/27] We supported **$S^2$-Attn** proposed by [LongLoRA](https://github.com/dvlab-research/LongLoRA) for the LLaMA models. Try `--shift_attn` argument to enable shift short attention.
|
||||
|
Loading…
x
Reference in New Issue
Block a user