Monitor Services

HPE Machine Learning Inferencing Software automatically configures monitoring of all deployed AI inference services. The types of information collected are:

tip icon Managing Observability

Inference Service Labels

Your inference service can be identified via one of its labels.

Label NameValueDescription
serving.kserve.io/inferenceserviceThe deployment name.Selects all instances of all versions of your inference service. Selectable in the Deployment Dashboard via the Deployment Name dropdown.
inference/packaged-modelThe packaged model name and version. For example: fb125m-model.v1Selectable in the Deployment Dashboard via the Packaged Model Version dropdown. By default, all versions of the deployment are shown.
inference/deployment-idThe deployment’s id value.For advanced use. Normally serving.kserve.io/inferenceservice is used as long as deployment names are not reused for different instances.
inference/packaged-model-idThe packaged model’s id value.For advanced use. Normally inference/packaged-model is used as long as packaged model names are not reused for different instances.
note icon Label Names in Grafana
These labels may be re-formatted by Grafana to use underscores instead of slashes (/), hyphens (-), and periods (.), for example, inference/packaged-model becomes inference_packaged_model.

Log Container Filters

You can filter logs by container type when accessing the Loki data source.

Container TypeDescription
kserve-containerLogs specifically from the running inference service code (bento)
aioli-loggerRequest/response logs events for the inference service
queue-proxyInternal kserve component supporting scale up/down
agentInternal kserve component tracking requests

Filtering Containers

Filter-out a container such as queue-proxy with the option container!="queue-proxy".

Viewing a Specific Container

View only a specific container such as your image output with container="kserve-container".