vllm.entrypoints.openai.completion.api_router ¶
ENDPOINT_LOAD_METRICS_FORMAT_HEADER_LABEL module-attribute ¶
attach_router ¶
completion ¶
completion(
request: Request,
) -> OpenAIServingCompletion | None
create_completion async ¶
create_completion(
request: CompletionRequest, raw_request: Request
)
Source code in vllm/entrypoints/openai/completion/api_router.py
render_completion async ¶
render_completion(
request: CompletionRequest, raw_request: Request
)
render completion request and return engine prompts without generating.