Monitor Services #
HPE Machine Learning Inferencing Software automatically configures monitoring of all deployed AI inference services. The types of information collected are:
Managing Observability
Inference Service Labels #
Your inference service can be identified via one of its labels.
Label Name | Value | Description |
---|---|---|
serving.kserve.io/inferenceservice | The deployment name. | Selects all instances of all versions of your inference service. Selectable in the Deployment Dashboard via the Deployment Name dropdown. |
inference/packaged-model | The packaged model name and version. For example: fb125m-model.v1 | Selectable in the Deployment Dashboard via the Packaged Model Version dropdown. By default, all versions of the deployment are shown. |
inference/deployment-id | The deployment’s id value. | For advanced use. Normally serving.kserve.io/inferenceservice is used as long as deployment names are not reused for different instances. |
inference/packaged-model-id | The packaged model’s id value. | For advanced use. Normally inference/packaged-model is used as long as packaged model names are not reused for different instances. |
Label Names in Grafana
/
), hyphens (-
), and periods (.
), for example, inference/packaged-model
becomes inference_packaged_model
.Log Container Filters #
You can filter logs by container type when accessing the Loki
data source.
Container Type | Description |
---|---|
kserve-container | Logs specifically from the running inference service code (bento) |
aioli-logger | Request/response logs events for the inference service |
queue-proxy | Internal kserve component supporting scale up/down |
agent | Internal kserve component tracking requests |
Filtering Containers #
Filter-out a container such as queue-proxy with the option container!="queue-proxy"
.
Viewing a Specific Container #
View only a specific container such as your image output with container="kserve-container"
.