mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 20:52:59 +08:00
parent
6c8bdeb6e0
commit
f6a8069159
@ -368,8 +368,7 @@ python src/export_model.py \
|
|||||||
--template default \
|
--template default \
|
||||||
--finetuning_type lora \
|
--finetuning_type lora \
|
||||||
--checkpoint_dir path_to_checkpoint \
|
--checkpoint_dir path_to_checkpoint \
|
||||||
--export_dir path_to_export \
|
--export_dir path_to_export
|
||||||
--fp16
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### API Demo
|
### API Demo
|
||||||
|
@ -367,8 +367,7 @@ python src/export_model.py \
|
|||||||
--template default \
|
--template default \
|
||||||
--finetuning_type lora \
|
--finetuning_type lora \
|
||||||
--checkpoint_dir path_to_checkpoint \
|
--checkpoint_dir path_to_checkpoint \
|
||||||
--output_dir path_to_export \
|
--export_dir path_to_export
|
||||||
--fp16
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### API 服务
|
### API 服务
|
||||||
|
@ -240,4 +240,7 @@ def load_model_and_tokenizer(
|
|||||||
trainable_params, all_param, 100 * trainable_params / all_param
|
trainable_params, all_param, 100 * trainable_params / all_param
|
||||||
))
|
))
|
||||||
|
|
||||||
|
if not is_trainable:
|
||||||
|
logger.info("This IS expected that the trainable params is 0 if you are using model for inference only.")
|
||||||
|
|
||||||
return model, tokenizer
|
return model, tokenizer
|
||||||
|
Loading…
x
Reference in New Issue
Block a user