From 769d81a83df00a7a2f1924f1fd7790c741237b11 Mon Sep 17 00:00:00 2001 From: Tsumugii24 <2792474059@qq.com> Date: Mon, 25 Mar 2024 22:31:03 +0800 Subject: [PATCH 1/3] Update README_zh.md Former-commit-id: deec57ec009ef6c08a90ad8e5800d6d5a936b337 --- README_zh.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README_zh.md b/README_zh.md index e359d960..71253c9a 100644 --- a/README_zh.md +++ b/README_zh.md @@ -308,10 +308,10 @@ cd LLaMA-Factory pip install -r requirements.txt ``` -如果要在 Windows 平台上开启量化 LoRA(QLoRA),需要安装预编译的 `bitsandbytes` 库, 支持 CUDA 11.1 到 12.2。 +如果要在 Windows 平台上开启量化 LoRA(QLoRA),需要安装预编译的 `bitsandbytes` 库, 支持 CUDA 11.1 到 12.2, 请根据 CUDA 版本的情况选择适合的 `release` 版本 [bitsandbytes-windows-webui]([Release Wheels · jllllll/bitsandbytes-windows-webui (github.com)](https://github.com/jllllll/bitsandbytes-windows-webui/releases/tag/wheels)) ```bash -pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.40.0-py3-none-win_amd64.whl +pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl ``` 如果要在 Windows 平台上开启 FlashAttention-2,需要安装预编译的 `flash-attn` 库,支持 CUDA 12.1 到 12.2,请根据需求到 [flash-attention](https://github.com/bdashore3/flash-attention/releases) 下载对应版本安装。 From 0c6a94e66df25a7e350804215e269c8bb52f6abc Mon Sep 17 00:00:00 2001 From: Tsumugii24 <2792474059@qq.com> Date: Mon, 25 Mar 2024 22:54:26 +0800 Subject: [PATCH 2/3] Update README_zh.md Former-commit-id: 34141ee0515c3e765ca0cb82a0625fb0abfba6f9 --- README_zh.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README_zh.md b/README_zh.md index 71253c9a..dc066e97 100644 --- a/README_zh.md +++ b/README_zh.md @@ -308,7 +308,7 @@ cd LLaMA-Factory pip install -r requirements.txt ``` -如果要在 Windows 平台上开启量化 LoRA(QLoRA),需要安装预编译的 `bitsandbytes` 库, 支持 CUDA 11.1 到 12.2, 请根据 CUDA 版本的情况选择适合的 `release` 版本 [bitsandbytes-windows-webui]([Release Wheels · jllllll/bitsandbytes-windows-webui (github.com)](https://github.com/jllllll/bitsandbytes-windows-webui/releases/tag/wheels)) +如果要在 Windows 平台上开启量化 LoRA(QLoRA),需要安装预编译的 `bitsandbytes` 库, 支持 CUDA 11.1 到 12.2, 请根据您使用的 CUDA 版本情况选择适合的 [bitsandbytes-windows-webui]([Release Wheels · jllllll/bitsandbytes-windows-webui (github.com)](https://github.com/jllllll/bitsandbytes-windows-webui/releases/tag/wheels)) `release` 版本。 ```bash pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl From 02665cd42b55f3f4f92f2bd7b0472ec2aeb9b76a Mon Sep 17 00:00:00 2001 From: Tsumugii24 <2792474059@qq.com> Date: Mon, 25 Mar 2024 22:54:38 +0800 Subject: [PATCH 3/3] Update README.md Former-commit-id: fd28fff2b9dfdb3e59b160c5fcee9cdc69e53564 --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index ea5d79bb..047633d9 100644 --- a/README.md +++ b/README.md @@ -308,10 +308,10 @@ cd LLaMA-Factory pip install -r requirements.txt ``` -If you want to enable the quantized LoRA (QLoRA) on the Windows platform, you will be required to install a pre-built version of `bitsandbytes` library, which supports CUDA 11.1 to 12.2. +If you want to enable the quantized LoRA (QLoRA) on the Windows platform, you will be required to install a pre-built version of `bitsandbytes` library, which supports CUDA 11.1 to 12.2, please select the appropriate [bitsandbytes-windows-webui]([Release Wheels · jllllll/bitsandbytes-windows-webui (github.com)](https://github.com/jllllll/bitsandbytes-windows-webui/releases/tag/wheels)) `release` version based on your CUDA version. ```bash -pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.40.0-py3-none-win_amd64.whl +pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl ``` To enable FlashAttention-2 on the Windows platform, you need to install the precompiled `flash-attn` library, which supports CUDA 12.1 to 12.2. Please download the corresponding version from [flash-attention](https://github.com/bdashore3/flash-attention/releases) based on your requirements.