From d65b2332cf5fa95be8a03ee9e4362c6e116a34f1 Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Tue, 7 May 2024 00:27:56 +0800 Subject: [PATCH] Update generating_args.py Former-commit-id: 7ae7ae64f0e9e8661f9efd30997f8b96673d467a --- src/llmtuner/hparams/generating_args.py | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/src/llmtuner/hparams/generating_args.py b/src/llmtuner/hparams/generating_args.py index 03e760e7..e3e196e9 100644 --- a/src/llmtuner/hparams/generating_args.py +++ b/src/llmtuner/hparams/generating_args.py @@ -1,5 +1,5 @@ from dataclasses import asdict, dataclass, field -from typing import Any, Dict, Union, Optional, List +from typing import Any, Dict @dataclass @@ -46,10 +46,6 @@ class GeneratingArguments: default=1.0, metadata={"help": "Exponential penalty to the length that is used with beam-based generation."}, ) - stop: Union[Optional[str], List[str]] = field( - default=None, - metadata={"help": "List of strings or string that stop the generation when they are generated. The returned output will not contain the stop strings."}, - ) def to_dict(self) -> Dict[str, Any]: args = asdict(self) if args.get("max_new_tokens", -1) > 0: