mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-15 08:08:09 +08:00
fix paligemma sft
Former-commit-id: 60682d04414be37e611d6470618a8d599703942b
This commit is contained in:
parent
09e78272c2
commit
2b65f8bd5c
@ -89,7 +89,7 @@ def preprocess_supervised_dataset(
|
||||
if processor is not None and hasattr(processor, "image_seq_length"): # paligemma case
|
||||
image_token_id = tokenizer.convert_tokens_to_ids(IMAGE_TOKEN)
|
||||
input_ids += [image_token_id] * getattr(processor, "image_seq_length")
|
||||
labels += [image_token_id] * getattr(processor, "image_seq_length")
|
||||
labels += [IGNORE_INDEX] * getattr(processor, "image_seq_length")
|
||||
|
||||
for turn_idx, (source_ids, target_ids) in enumerate(
|
||||
template.encode_multiturn(
|
||||
|
Loading…
x
Reference in New Issue
Block a user