[misc] move wechat out (#9223)

This commit is contained in:
Yaowei Zheng 2025-10-02 02:06:09 +08:00 committed by GitHub
parent 7dd910f067
commit bcc2c1fd8f
4 changed files with 7 additions and 7 deletions

View File

@ -36,7 +36,7 @@
</div>
👋 Join our [WeChat](assets/wechat/main.jpg), [NPU](assets/wechat/npu.jpg), [Lab4AI](assets/wechat/lab4ai.jpg), [LLaMA Factory Online](assets/wechat/online.jpg) user group.
👋 Join our [WeChat](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg), [NPU](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg), [Lab4AI](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg), [LLaMA Factory Online](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.jpg) user group.
\[ English | [中文](README_zh.md) \]

View File

@ -36,7 +36,7 @@
</div>
👋 加入我们的[微信群](assets/wechat/main.jpg)、[NPU 用户群](assets/wechat/npu.jpg)、[大模型实验室群](assets/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](assets/wechat/online.png)。
👋 加入我们的[微信群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg)、[NPU 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg)、[大模型实验室群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.png)。
\[ [English](README.md) | 中文 \]

View File

@ -6,7 +6,7 @@
model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
trust_remote_code: true
### Method configuration
### Method configuration
stage: sft
do_train: true
finetuning_type: full
@ -40,9 +40,9 @@ fp8: true
fp8_backend: torchao # Use TorchAO backend for FP8
fp8_enable_fsdp_float8_all_gather: false # Not used with DeepSpeed
### DeepSpeed configuration
### DeepSpeed configuration
deepspeed: examples/deepspeed/ds_z3_fp8_config.json
### Logging configuration
report_to: wandb
run_name: llama3_fp8_deepspeed_sft
run_name: llama3_fp8_deepspeed_sft

View File

@ -1,4 +1,4 @@
# FP8 training example with FSDP
# FP8 training example with FSDP
# This config demonstrates FP8 mixed precision training using HuggingFace Accelerate
# with FSDP for distributed training and float8 all-gather optimization
@ -48,4 +48,4 @@ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
### Logging configuration
report_to: wandb
run_name: llama3_fp8_fsdp_sft
run_name: llama3_fp8_fsdp_sft