From 116de2ce4813f47ae5b30810e7a70215653f4d25 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Sun, 25 Feb 2024 16:26:08 +0800 Subject: [PATCH] update readme Former-commit-id: 261f631a1cc708a5713e40102ad558c5dfa6a379 --- README.md | 8 ++++---- README_zh.md | 8 ++++---- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/README.md b/README.md index 59dd0690..358d775a 100644 --- a/README.md +++ b/README.md @@ -516,7 +516,7 @@ python src/export_model.py \ > [!TIP] > Use `--export_quantization_bit 4` and `--export_quantization_dataset data/c4_demo.json` to quantize the model after merging the LoRA weights. -### API Demo +### Inference with OpenAI-style API ```bash python src/api_demo.py \ @@ -529,7 +529,7 @@ python src/api_demo.py \ > [!TIP] > Visit `http://localhost:8000/docs` for API documentation. -### CLI Demo +### Inference with command line ```bash python src/cli_demo.py \ @@ -539,7 +539,7 @@ python src/cli_demo.py \ --finetuning_type lora ``` -### Web Demo +### Inference with web browser ```bash python src/web_demo.py \ @@ -576,7 +576,7 @@ CUDA_VISIBLE_DEVICES=0 python src/train_bash.py \ --template default \ --finetuning_type lora \ --output_dir path_to_predict_result \ - --per_device_eval_batch_size 8 \ + --per_device_eval_batch_size 1 \ --max_samples 100 \ --predict_with_generate \ --fp16 diff --git a/README_zh.md b/README_zh.md index 6a914541..41299bfb 100644 --- a/README_zh.md +++ b/README_zh.md @@ -516,7 +516,7 @@ python src/export_model.py \ > [!TIP] > 合并 LoRA 权重之后可再次使用 `--export_quantization_bit 4` 和 `--export_quantization_dataset data/c4_demo.json` 量化模型。 -### API 服务 +### 使用 OpenAI 风格 API 推理 ```bash python src/api_demo.py \ @@ -529,7 +529,7 @@ python src/api_demo.py \ > [!TIP] > 关于 API 文档请见 `http://localhost:8000/docs`。 -### 命令行测试 +### 使用命令行推理 ```bash python src/cli_demo.py \ @@ -539,7 +539,7 @@ python src/cli_demo.py \ --finetuning_type lora ``` -### 浏览器测试 +### 使用浏览器推理 ```bash python src/web_demo.py \ @@ -576,7 +576,7 @@ CUDA_VISIBLE_DEVICES=0 python src/train_bash.py \ --template default \ --finetuning_type lora \ --output_dir path_to_predict_result \ - --per_device_eval_batch_size 8 \ + --per_device_eval_batch_size 1 \ --max_samples 100 \ --predict_with_generate \ --fp16