Monitor Services

HPE Machine Learning Inferencing Software automatically configures monitoring of all deployed AI inference services. The types of information collected are:

Managing Observability

Inference Service Labels

Your inference service can be identified via one of its labels.

Label Name Value Description
serving.kserve.io/inferenceservice The deployment name. Selects all instances of all versions of your inference service. Selectable in the Deployment Dashboard via the Deployment Name dropdown.
inference/packaged-model The packaged model name and version. For example: fb125m-model.v1 Selectable in the Deployment Dashboard via the Packaged Model Version dropdown. By default, all versions of the deployment are shown.
inference/deployment-id The deployment’s id value. For advanced use. Normally serving.kserve.io/inferenceservice is used as long as deployment names are not reused for different instances.
inference/packaged-model-id The packaged model’s id value. For advanced use. Normally inference/packaged-model is used as long as packaged model names are not reused for different instances.
Label Names in Grafana
These labels may be re-formatted by Grafana to use underscores instead of slashes (/), hyphens (-), and periods (.), for example, inference/packaged-model becomes inference_packaged_model.

Log Container Filters

You can filter logs by container type when accessing the Loki data source.

Container Type Description
kserve-container Logs specifically from the running inference service code (bento)
aioli-logger Request/response logs events for the inference service
queue-proxy Internal kserve component supporting scale up/down
agent Internal kserve component tracking requests

Filtering Containers

Filter-out a container such as queue-proxy with the option container!="queue-proxy".

Viewing a Specific Container

View only a specific container such as your image output with container="kserve-container".