mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 22:32:54 +08:00
fix api
Former-commit-id: 8333795fcb25728dc7146596869f06e5e9533e58
This commit is contained in:
parent
38109c3da3
commit
f1010ff65f
@ -5,11 +5,14 @@
|
|||||||
|
|
||||||
import uvicorn
|
import uvicorn
|
||||||
|
|
||||||
|
from llmtuner import ChatModel
|
||||||
from llmtuner.api.app import create_app
|
from llmtuner.api.app import create_app
|
||||||
|
from llmtuner.tuner import get_infer_args
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
app = create_app()
|
chat_model = ChatModel(*get_infer_args())
|
||||||
|
app = create_app(chat_model)
|
||||||
uvicorn.run(app, host="0.0.0.0", port=8000, workers=1)
|
uvicorn.run(app, host="0.0.0.0", port=8000, workers=1)
|
||||||
|
|
||||||
|
|
||||||
|
@ -30,7 +30,7 @@ async def lifespan(app: FastAPI): # collects GPU memory
|
|||||||
torch_gc()
|
torch_gc()
|
||||||
|
|
||||||
|
|
||||||
def create_app(chat_model: ChatModel):
|
def create_app(chat_model: ChatModel) -> FastAPI:
|
||||||
app = FastAPI(lifespan=lifespan)
|
app = FastAPI(lifespan=lifespan)
|
||||||
|
|
||||||
app.add_middleware(
|
app.add_middleware(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user