mirror of
				https://github.com/hiyouga/LLaMA-Factory.git
				synced 2025-11-04 18:02:19 +08:00 
			
		
		
		
	[misc] move wechat out (#9223)
This commit is contained in:
		
							parent
							
								
									2a797e2562
								
							
						
					
					
						commit
						6f743571b1
					
				@ -36,7 +36,7 @@
 | 
			
		||||
 | 
			
		||||
</div>
 | 
			
		||||
 | 
			
		||||
👋 Join our [WeChat](assets/wechat/main.jpg), [NPU](assets/wechat/npu.jpg), [Lab4AI](assets/wechat/lab4ai.jpg), [LLaMA Factory Online](assets/wechat/online.jpg) user group.
 | 
			
		||||
👋 Join our [WeChat](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg), [NPU](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg), [Lab4AI](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg), [LLaMA Factory Online](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.jpg) user group.
 | 
			
		||||
 | 
			
		||||
\[ English | [中文](README_zh.md) \]
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -36,7 +36,7 @@
 | 
			
		||||
 | 
			
		||||
</div>
 | 
			
		||||
 | 
			
		||||
👋 加入我们的[微信群](assets/wechat/main.jpg)、[NPU 用户群](assets/wechat/npu.jpg)、[大模型实验室群](assets/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](assets/wechat/online.png)。
 | 
			
		||||
👋 加入我们的[微信群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/main.jpg)、[NPU 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/npu.jpg)、[大模型实验室群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/lab4ai.jpg) 或 [LLaMA Factory Online 用户群](https://github.com/hiyouga/llamafactory-community/blob/main/wechat/online.png)。
 | 
			
		||||
 | 
			
		||||
\[ [English](README.md) | 中文 \]
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
										
											Binary file not shown.
										
									
								
							| 
		 Before Width: | Height: | Size: 74 KiB  | 
										
											Binary file not shown.
										
									
								
							| 
		 Before Width: | Height: | Size: 168 KiB  | 
										
											Binary file not shown.
										
									
								
							| 
		 Before Width: | Height: | Size: 167 KiB  | 
										
											Binary file not shown.
										
									
								
							| 
		 Before Width: | Height: | Size: 128 KiB  | 
@ -42,4 +42,4 @@
 | 
			
		||||
      "enabled": false
 | 
			
		||||
    }
 | 
			
		||||
  }
 | 
			
		||||
}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@ -6,7 +6,7 @@
 | 
			
		||||
model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
 | 
			
		||||
trust_remote_code: true
 | 
			
		||||
 | 
			
		||||
### Method configuration  
 | 
			
		||||
### Method configuration
 | 
			
		||||
stage: sft
 | 
			
		||||
do_train: true
 | 
			
		||||
finetuning_type: full
 | 
			
		||||
@ -40,9 +40,9 @@ fp8: true
 | 
			
		||||
fp8_backend: torchao  # Use TorchAO backend for FP8
 | 
			
		||||
fp8_enable_fsdp_float8_all_gather: false  # Not used with DeepSpeed
 | 
			
		||||
 | 
			
		||||
### DeepSpeed configuration  
 | 
			
		||||
### DeepSpeed configuration
 | 
			
		||||
deepspeed: examples/deepspeed/ds_z3_fp8_config.json
 | 
			
		||||
 | 
			
		||||
### Logging configuration
 | 
			
		||||
report_to: wandb
 | 
			
		||||
run_name: llama3_fp8_deepspeed_sft
 | 
			
		||||
run_name: llama3_fp8_deepspeed_sft
 | 
			
		||||
 | 
			
		||||
@ -1,4 +1,4 @@
 | 
			
		||||
# FP8 training example with FSDP 
 | 
			
		||||
# FP8 training example with FSDP
 | 
			
		||||
# This config demonstrates FP8 mixed precision training using HuggingFace Accelerate
 | 
			
		||||
# with FSDP for distributed training and float8 all-gather optimization
 | 
			
		||||
 | 
			
		||||
@ -48,4 +48,4 @@ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
 | 
			
		||||
 | 
			
		||||
### Logging configuration
 | 
			
		||||
report_to: wandb
 | 
			
		||||
run_name: llama3_fp8_fsdp_sft
 | 
			
		||||
run_name: llama3_fp8_fsdp_sft
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user