mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-11-05 02:12:14 +08:00
[misc] move wechat out (#9223)
This commit is contained in:
parent
2a797e2562
commit
6f743571b1
@ -36,7 +36,7 @@
|
|||||||
|
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
👋 Join our [WeChat](assets/wechat/main.jpg), [NPU](assets/wechat/npu.jpg), [Lab4AI](assets/wechat/lab4ai.jpg), [LLaMA Factory Online](assets/wechat/online.jpg) user group.
|
👋 Join our [WeChat](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg), [NPU](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg), [Lab4AI](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg), [LLaMA Factory Online](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.jpg) user group.
|
||||||
|
|
||||||
\[ English | [中文](README_zh.md) \]
|
\[ English | [中文](README_zh.md) \]
|
||||||
|
|
||||||
|
|||||||
@ -36,7 +36,7 @@
|
|||||||
|
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
👋 加入我们的[微信群](assets/wechat/main.jpg)、[NPU 用户群](assets/wechat/npu.jpg)、[大模型实验室群](assets/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](assets/wechat/online.png)。
|
👋 加入我们的[微信群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg)、[NPU 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg)、[大模型实验室群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.png)。
|
||||||
|
|
||||||
\[ [English](README.md) | 中文 \]
|
\[ [English](README.md) | 中文 \]
|
||||||
|
|
||||||
|
|||||||
Binary file not shown.
|
Before Width: | Height: | Size: 74 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 168 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 167 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 128 KiB |
@ -42,4 +42,4 @@
|
|||||||
"enabled": false
|
"enabled": false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -6,7 +6,7 @@
|
|||||||
model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
|
model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
|
||||||
trust_remote_code: true
|
trust_remote_code: true
|
||||||
|
|
||||||
### Method configuration
|
### Method configuration
|
||||||
stage: sft
|
stage: sft
|
||||||
do_train: true
|
do_train: true
|
||||||
finetuning_type: full
|
finetuning_type: full
|
||||||
@ -40,9 +40,9 @@ fp8: true
|
|||||||
fp8_backend: torchao # Use TorchAO backend for FP8
|
fp8_backend: torchao # Use TorchAO backend for FP8
|
||||||
fp8_enable_fsdp_float8_all_gather: false # Not used with DeepSpeed
|
fp8_enable_fsdp_float8_all_gather: false # Not used with DeepSpeed
|
||||||
|
|
||||||
### DeepSpeed configuration
|
### DeepSpeed configuration
|
||||||
deepspeed: examples/deepspeed/ds_z3_fp8_config.json
|
deepspeed: examples/deepspeed/ds_z3_fp8_config.json
|
||||||
|
|
||||||
### Logging configuration
|
### Logging configuration
|
||||||
report_to: wandb
|
report_to: wandb
|
||||||
run_name: llama3_fp8_deepspeed_sft
|
run_name: llama3_fp8_deepspeed_sft
|
||||||
|
|||||||
@ -1,4 +1,4 @@
|
|||||||
# FP8 training example with FSDP
|
# FP8 training example with FSDP
|
||||||
# This config demonstrates FP8 mixed precision training using HuggingFace Accelerate
|
# This config demonstrates FP8 mixed precision training using HuggingFace Accelerate
|
||||||
# with FSDP for distributed training and float8 all-gather optimization
|
# with FSDP for distributed training and float8 all-gather optimization
|
||||||
|
|
||||||
@ -48,4 +48,4 @@ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
|
|||||||
|
|
||||||
### Logging configuration
|
### Logging configuration
|
||||||
report_to: wandb
|
report_to: wandb
|
||||||
run_name: llama3_fp8_fsdp_sft
|
run_name: llama3_fp8_fsdp_sft
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user