mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-15 11:20:35 +08:00
support quantization in export model
This commit is contained in:
@@ -479,7 +479,10 @@ python src/export_model.py \
|
||||
```
|
||||
|
||||
> [!WARNING]
|
||||
> Merging LoRA weights into a GPTQ quantized model is not supported.
|
||||
> Merging LoRA weights into a quantized model is not supported.
|
||||
|
||||
> [!TIP]
|
||||
> Use `--export_quantization_bit 4` and `--export_quantization_dataset data/wiki_demo.txt` to quantize the model.
|
||||
|
||||
### API Demo
|
||||
|
||||
|
||||
Reference in New Issue
Block a user