diff --git a/README.md b/README.md index 34d53d7d..e3842435 100644 --- a/README.md +++ b/README.md @@ -558,9 +558,9 @@ python src/export_model.py \ > Merging LoRA weights into a quantized model is not supported. > [!TIP] -> Use `--model_name_or_path path_to_export` only to use the exported model. +> Use `--model_name_or_path path_to_export` solely to use the exported model. > -> Use `--export_quantization_bit 4` and `--export_quantization_dataset data/c4_demo.json` to quantize the model after merging the LoRA weights. +> Use `--export_quantization_bit 4` and `--export_quantization_dataset data/c4_demo.json` to quantize the model with AutoGPTQ after merging the LoRA weights. ### Inference with OpenAI-style API diff --git a/README_zh.md b/README_zh.md index e1917bdb..498b2d5b 100644 --- a/README_zh.md +++ b/README_zh.md @@ -559,7 +559,7 @@ python src/export_model.py \ > [!TIP] > 仅使用 `--model_name_or_path path_to_export` 来加载导出后的模型。 > -> 合并 LoRA 权重之后可再次使用 `--export_quantization_bit 4` 和 `--export_quantization_dataset data/c4_demo.json` 量化模型。 +> 合并 LoRA 权重之后可再次使用 `--export_quantization_bit 4` 和 `--export_quantization_dataset data/c4_demo.json` 基于 AutoGPTQ 量化模型。 ### 使用 OpenAI 风格 API 推理