mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 06:12:50 +08:00
Release v0.1.7
Former-commit-id: 9c9009f49fdff83a29b83d8f97eb5c99e2574256
This commit is contained in:
parent
9f1688924d
commit
acaac6df9e
@ -6,4 +6,4 @@ from llmtuner.tuner import export_model, run_exp
|
|||||||
from llmtuner.webui import create_ui, create_web_demo
|
from llmtuner.webui import create_ui, create_web_demo
|
||||||
|
|
||||||
|
|
||||||
__version__ = "0.1.6"
|
__version__ = "0.1.7"
|
||||||
|
@ -55,7 +55,7 @@ class Seq2SeqPeftTrainer(PeftTrainer):
|
|||||||
self.tokenizer.pad_token_id * torch.ones_like(generated_tokens[:, :max(prompt_len, label_len)])
|
self.tokenizer.pad_token_id * torch.ones_like(generated_tokens[:, :max(prompt_len, label_len)])
|
||||||
)
|
)
|
||||||
|
|
||||||
return (loss, generated_tokens, labels)
|
return loss, generated_tokens, labels
|
||||||
|
|
||||||
def _pad_tensors_to_target_len(
|
def _pad_tensors_to_target_len(
|
||||||
self,
|
self,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user