mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 14:22:51 +08:00
Merge pull request #5973 from JJJJerry/fix_vllm_generate
fix VllmEngine: 将inputs参数替换为prompt Former-commit-id: 40a2fcc02dad2f1633c117d457651b221b8c5ae0
This commit is contained in:
commit
0386fa6a4f
@ -173,7 +173,7 @@ class VllmEngine(BaseEngine):
|
|||||||
multi_modal_data = None
|
multi_modal_data = None
|
||||||
|
|
||||||
result_generator = self.model.generate(
|
result_generator = self.model.generate(
|
||||||
inputs={"prompt_token_ids": prompt_ids, "multi_modal_data": multi_modal_data},
|
{"prompt_token_ids": prompt_ids, "multi_modal_data": multi_modal_data},
|
||||||
sampling_params=sampling_params,
|
sampling_params=sampling_params,
|
||||||
request_id=request_id,
|
request_id=request_id,
|
||||||
lora_request=self.lora_request,
|
lora_request=self.lora_request,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user