mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-15 03:10:35 +08:00
tiny fix
This commit is contained in:
@@ -41,7 +41,7 @@ def prepare_4d_attention_mask(attention_mask_with_indices: "torch.Tensor", dtype
|
|||||||
[x, x, o, x, x, x],
|
[x, x, o, x, x, x],
|
||||||
[x, x, o, o, x, x],
|
[x, x, o, o, x, x],
|
||||||
[x, x, o, o, o, x],
|
[x, x, o, o, o, x],
|
||||||
[x, x, o, x, x, x],
|
[x, x, x, x, x, x],
|
||||||
]
|
]
|
||||||
]
|
]
|
||||||
]
|
]
|
||||||
|
|||||||
Reference in New Issue
Block a user