diff --git a/src/llmtuner/chat/chat_model.py b/src/llmtuner/chat/chat_model.py index d92848c2..1a8f95aa 100644 --- a/src/llmtuner/chat/chat_model.py +++ b/src/llmtuner/chat/chat_model.py @@ -37,9 +37,9 @@ class ChatModel: tools: Optional[str] = None, **input_kwargs, ) -> Tuple[Dict[str, Any], int]: - messages += [{"role": "assistant", "content": ""}] + new_messages = messages + [{"role": "assistant", "content": ""}] prompt, _ = self.template.encode_oneturn( - tokenizer=self.tokenizer, messages=messages, system=system, tools=tools + tokenizer=self.tokenizer, messages=new_messages, system=system, tools=tools ) prompt_length = len(prompt) input_ids = torch.tensor([prompt], device=self.model.device) diff --git a/src/llmtuner/data/template.py b/src/llmtuner/data/template.py index d4fd88fc..a000a7f9 100644 --- a/src/llmtuner/data/template.py +++ b/src/llmtuner/data/template.py @@ -416,7 +416,7 @@ register_template( "by the user such as English and 中文." ), stop_words=["<|im_end|>"], - replace_eos=True, + efficient_eos=True, )