mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-25 15:22:48 +08:00
The training efficiency of the Ascend 910A has been significantly enhanced, leveraging the full computational power of the NPU (Neural Processing Unit) and the capabilities of torch_npu, a PyTorch library optimized for NPUs. This improvement has resulted in a remarkable tenfold increase in efficiency.
Former-commit-id: 28ae947161d4670d4f865cbaad84397d47215a53
This commit is contained in:
parent
2a53a43ac7
commit
7b0629dac4
@ -1,3 +1,7 @@
|
|||||||
|
import os
|
||||||
|
import torch
|
||||||
|
import torch_npu
|
||||||
|
from torch_npu.contrib import transfer_to_npu
|
||||||
from llmtuner.train.tuner import run_exp
|
from llmtuner.train.tuner import run_exp
|
||||||
|
|
||||||
|
|
||||||
@ -11,4 +15,6 @@ def _mp_fn(index):
|
|||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
use_jit_compile = os.getenv('JIT_COMPILE', 'False').lower() in ['true', '1']
|
||||||
|
torch.npu.set_compile_mode(jit_compile=use_jit_compile)
|
||||||
main()
|
main()
|
||||||
|
Loading…
x
Reference in New Issue
Block a user