mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2026-02-26 15:56:00 +08:00
44 lines
789 B
YAML
44 lines
789 B
YAML
model: Qwen/Qwen3-0.6B
|
|
trust_remote_code: true
|
|
model_class: llm
|
|
|
|
template: qwen3_nothink
|
|
|
|
# PEFT Configuration
|
|
peft_config:
|
|
name: lora
|
|
r: 16
|
|
lora_alpha: 32
|
|
lora_dropout: 0.05
|
|
target_modules: all
|
|
|
|
# Kernel Config
|
|
kernel_config:
|
|
name: auto
|
|
include_kernels: auto
|
|
|
|
# FSDP Config
|
|
dist_config:
|
|
name: fsdp2
|
|
dcp_path: null
|
|
|
|
# Quantization Config
|
|
quant_config:
|
|
name: bnb # choice: auto/bnb if auto is selected, the quantization method will be automatically selected based on the model and environment.
|
|
quantization_bit: 4 # choice: 8/4(bnb)
|
|
|
|
### data
|
|
train_dataset: data/v1_sft_demo.yaml
|
|
|
|
### training
|
|
output_dir: outputs/test_quantization
|
|
micro_batch_size: 1
|
|
cutoff_len: 2048
|
|
learning_rate: 1.0e-4
|
|
bf16: false
|
|
max_steps: 10
|
|
|
|
### sample
|
|
sample_backend: hf
|
|
max_new_tokens: 128
|