diff --git a/assets/wechat.jpg b/assets/wechat.jpg index 79cdc21a..d933efed 100644 Binary files a/assets/wechat.jpg and b/assets/wechat.jpg differ diff --git a/assets/wechat_npu.jpg b/assets/wechat_npu.jpg index 5104d61c..60c311a9 100644 Binary files a/assets/wechat_npu.jpg and b/assets/wechat_npu.jpg differ diff --git a/examples/inference/llama3_full_sft.yaml b/examples/inference/llama3_full_sft.yaml new file mode 100644 index 00000000..d0c33209 --- /dev/null +++ b/examples/inference/llama3_full_sft.yaml @@ -0,0 +1,3 @@ +model_name_or_path: saves/llama3-8b/full/sft +template: llama3 +infer_backend: huggingface # choices: [huggingface, vllm] diff --git a/examples/inference/llama3_lora_sft.yaml b/examples/inference/llama3_lora_sft.yaml index ec5d8732..25d5457f 100644 --- a/examples/inference/llama3_lora_sft.yaml +++ b/examples/inference/llama3_lora_sft.yaml @@ -1,5 +1,4 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct adapter_name_or_path: saves/llama3-8b/lora/sft template: llama3 -finetuning_type: lora infer_backend: huggingface # choices: [huggingface, vllm]