mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 06:12:50 +08:00
Merge pull request #5990 from hiyouga/hiyouga/dev_vllm
[generate] fix vllm config args Former-commit-id: 3eebae892be74e6d9d4e7d1f6f1fae00080957f9
This commit is contained in:
commit
8ae2056b59
@ -83,7 +83,8 @@ class VllmEngine(BaseEngine):
|
||||
"enable_lora": model_args.adapter_name_or_path is not None,
|
||||
"max_lora_rank": model_args.vllm_max_lora_rank,
|
||||
}
|
||||
engine_args.update(model_args.vllm_config)
|
||||
if isinstance(model_args.vllm_config, dict):
|
||||
engine_args.update(model_args.vllm_config)
|
||||
|
||||
if getattr(config, "is_yi_vl_derived_model", None):
|
||||
import vllm.model_executor.models.llava
|
||||
|
Loading…
x
Reference in New Issue
Block a user