fix bug in DPO data collator

Former-commit-id: 90bd085ae4c2775f1e82e045ab0157a451774082
This commit is contained in:
hiyouga 2023-09-08 20:45:07 +08:00
parent 405df0f63d
commit 43a20c67d4

View File

@ -16,7 +16,7 @@ class DPODataCollatorWithPadding(DataCollatorForSeq2Seq):
if self.tokenizer.padding_side == "left":
start, end = feature.size(0) - answer_len, feature.size(0)
else:
start, end = prompt_len, answer_len
start, end = prompt_len, prompt_len + answer_len
padded_tensor = self.label_pad_token_id * torch.ones_like(feature)
padded_tensor[start:end] = feature[start:end]
padded_labels.append(padded_tensor)