mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 06:12:50 +08:00
Merge pull request #3449 from hiyouga/mllm
add webui backend option Former-commit-id: fcfbd8c3003989a5671f2a31a439ad2c336b55f1
This commit is contained in:
commit
317252f2bf
@ -31,7 +31,10 @@ class WebChatModel(ChatModel):
|
||||
if demo_mode and os.environ.get("DEMO_MODEL") and os.environ.get("DEMO_TEMPLATE"): # load demo model
|
||||
model_name_or_path = os.environ.get("DEMO_MODEL")
|
||||
template = os.environ.get("DEMO_TEMPLATE")
|
||||
super().__init__(dict(model_name_or_path=model_name_or_path, template=template))
|
||||
infer_backend = os.environ.get("DEMO_BACKEND", "huggingface")
|
||||
super().__init__(
|
||||
dict(model_name_or_path=model_name_or_path, template=template, infer_backend=infer_backend)
|
||||
)
|
||||
|
||||
@property
|
||||
def loaded(self) -> bool:
|
||||
|
Loading…
x
Reference in New Issue
Block a user