⚠ This page is served via a proxy. Original site: https://github.com
This service does not collect credentials or authentication data.
Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 5 additions & 2 deletions apipod/core/routers/router_mixins/_fast_api_llm_mixin.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
from fastapi.responses import StreamingResponse

from apipod.core.routers.router_mixins._base_llm_mixin import _BaseLLMMixin
from apipod.settings import SERVER_DOMAIN

class _FastAPILLMMixin(_BaseLLMMixin):
"""
Expand All @@ -20,7 +21,9 @@ async def handle_llm_request(self, func, openai_req, should_use_queue, res_model
job_function=func,
job_params={"payload": openai_req.dict()}
)
return JobResultFactory.from_base_job(job)

ret_job = JobResultFactory.from_base_job(job)
ret_job.refresh_job_url = f"{SERVER_DOMAIN}/status?job_id={ret_job.id}"
return ret_job

raw_res = await self._execute_func(func, payload=openai_req, **kwargs)
return self._wrap_llm_response(raw_res, res_model, endpoint_type, openai_req)