From 05277ee86437e0e53eff016985a19e7d205e03bb Mon Sep 17 00:00:00 2001 From: hiyouga <467089858@qq.com> Date: Wed, 29 May 2024 01:20:07 +0800 Subject: [PATCH] fix hf chat engine Former-commit-id: 97346c1d3d87f0bd5ddcd70ff485f6a8273244aa --- src/llamafactory/chat/hf_engine.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/src/llamafactory/chat/hf_engine.py b/src/llamafactory/chat/hf_engine.py index 2148f8cd..ad0e90fe 100644 --- a/src/llamafactory/chat/hf_engine.py +++ b/src/llamafactory/chat/hf_engine.py @@ -8,6 +8,7 @@ import torch from transformers import GenerationConfig, TextIteratorStreamer from ..data import get_template_and_fix_tokenizer +from ..extras.logging import get_logger from ..extras.misc import get_logits_processor from ..model import load_model, load_tokenizer from .base_engine import BaseEngine, Response @@ -23,6 +24,9 @@ if TYPE_CHECKING: from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments +logger = get_logger(__name__) + + class HuggingfaceEngine(BaseEngine): def __init__( self, @@ -92,7 +96,7 @@ class HuggingfaceEngine(BaseEngine): stop: Optional[Union[str, List[str]]] = input_kwargs.pop("stop", None) if stop is not None: - raise ValueError("Stop parameter is not supported in Huggingface engine yet.") + logger.warning("Stop parameter is not supported in Huggingface engine yet.") generating_args = generating_args.copy() generating_args.update(