Get desktop application:
View/edit binary Protocol Buffers messages
Service definition
Check health status of the inference server.
Ping / healthcheck response
(e.g. "ok" or "not_ok")
Get service information (version, release date etc.)
Service information repsonse
(e.g. "0.1.0")
(e.g. "cpu", "gpu", "local" etc)
List available models from Hugging Face Hub
/ Pre-load a model for inference
Get model information from the deployment
Run the inference request
Stream the inference response
Register shared memory
Unregister shared memory
Upload file
Delete file
Register system shared memory request
Used as request type in: InferenceService.DeleteFile, InferenceService.LoadModel, InferenceService.RegisterSystemSharedMemory, InferenceService.Run, InferenceService.Stream, InferenceService.UnregisterSystemSharedMemory, InferenceService.UploadFile
Register system shared memory response
Used as response type in: InferenceService.GetModelCatalog, InferenceService.GetModelInfo, InferenceService.ListModels, InferenceService.LoadModel, InferenceService.RegisterSystemSharedMemory, InferenceService.Run, InferenceService.Stream, InferenceService.UnregisterSystemSharedMemory, InferenceService.UploadFile