mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-22 22:02:51 +08:00
remove empty stream response
Former-commit-id: e984ba3167d765837cff1030acf59528bcde2f85
This commit is contained in:
parent
6672ad7a83
commit
87b9f70ab4
@ -156,9 +156,10 @@ async def create_stream_chat_completion_response(
|
|||||||
top_p=request.top_p,
|
top_p=request.top_p,
|
||||||
max_new_tokens=request.max_tokens,
|
max_new_tokens=request.max_tokens,
|
||||||
):
|
):
|
||||||
yield _create_stream_chat_completion_chunk(
|
if len(new_token) != 0:
|
||||||
completion_id=completion_id, model=request.model, delta=ChatCompletionMessage(content=new_token)
|
yield _create_stream_chat_completion_chunk(
|
||||||
)
|
completion_id=completion_id, model=request.model, delta=ChatCompletionMessage(content=new_token)
|
||||||
|
)
|
||||||
|
|
||||||
yield _create_stream_chat_completion_chunk(
|
yield _create_stream_chat_completion_chunk(
|
||||||
completion_id=completion_id, model=request.model, delta=ChatCompletionMessage(), finish_reason=Finish.STOP
|
completion_id=completion_id, model=request.model, delta=ChatCompletionMessage(), finish_reason=Finish.STOP
|
||||||
|
Loading…
x
Reference in New Issue
Block a user