nip.language_model_server.server.start_vllm_server#
- async nip.language_model_server.server.start_vllm_server(request: VllmStartRequest) VllmStartResponse[source]#
- Start the vLLM server with the specified model. - Parameters:
- request (VllmStartRequest) – A request containing the model name to be served by the vLLM server. 
- Returns:
- response (VllmStartResponse) – A data structure containing the success message, model name, and port on which the vLLM server is running. 
- Raises:
- HTTPException – If the vLLM server fails to start.