mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-02 19:52:50 +08:00
update examples
Former-commit-id: 4c00bcdcaeb675c9fdb3e977c27c3604d7895ae2
This commit is contained in:
parent
c561b268ef
commit
eb363b04b9
@ -1,7 +1,5 @@
|
||||
#!/bin/bash
|
||||
|
||||
pip install -e ../../../.
|
||||
|
||||
python ../../../scripts/llama_pro.py \
|
||||
--model_name_or_path meta-llama/Llama-2-7b-hf \
|
||||
--output_dir ../../../models/llama2-7b-pro \
|
||||
|
@ -10,6 +10,7 @@ CUDA_VISIBLE_DEVICES=0 python ../../../src/train_bash.py \
|
||||
--finetuning_type freeze \
|
||||
--name_module_trainable all \
|
||||
--num_layer_trainable 8 \
|
||||
--use_llama_pro \
|
||||
--output_dir ../../../saves/LLaMA2-7B-Pro/lora/sft \
|
||||
--overwrite_cache \
|
||||
--overwrite_output_dir \
|
||||
|
@ -33,5 +33,6 @@ python -m torch.distributed.run \
|
||||
--num_train_epochs 3.0 \
|
||||
--max_samples 3000 \
|
||||
--val_size 0.1 \
|
||||
--ddp_timeout 1800000 \
|
||||
--plot_loss \
|
||||
--fp16
|
||||
|
@ -27,5 +27,6 @@ deepspeed --num_gpus 4 ../../src/train_bash.py \
|
||||
--num_train_epochs 3.0 \
|
||||
--max_samples 3000 \
|
||||
--val_size 0.1 \
|
||||
--ddp_timeout 1800000 \
|
||||
--plot_loss \
|
||||
--fp16
|
||||
|
@ -30,5 +30,6 @@ CUDA_VISIBLE_DEVICES=0,1,2,3 accelerate launch \
|
||||
--num_train_epochs 3.0 \
|
||||
--max_samples 3000 \
|
||||
--val_size 0.1 \
|
||||
--ddp_timeout 1800000 \
|
||||
--plot_loss \
|
||||
--fp16
|
||||
|
@ -30,5 +30,6 @@ CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 accelerate launch \
|
||||
--num_train_epochs 3.0 \
|
||||
--max_samples 3000 \
|
||||
--val_size 0.1 \
|
||||
--ddp_timeout 1800000 \
|
||||
--plot_loss \
|
||||
--fp16
|
||||
|
Loading…
x
Reference in New Issue
Block a user