1.1.0
Welcome to the 1.1.0 release of HPE Machine Learning Inferencing Software (MLIS).
REST API Docs
http://<your-mlis-url>/docs/rest-api/
.Highlights #
This release includes the following features:
Deployments #
Deployment Tokens #
Create deployment tokens to control access to your inference service endpoint.
- UI Feature: Create and manage deployment tokens via the UI.
- CLI Feature: In addition to the timestamp format, you can manage the expiration date of an access token by applying a simple date-time or simple date format.
Model Management #
- Feature: Load models from Persistent Volume Claims (PVC) using
pvc://
model URL prefixes within the same Kubernetes namespace.
Admin #
- Feature: Automatically collect and report anonymous customer telemetry data to improve product quality and support.
Known Issues #
- DB Connections: TLS/SSL connections are not supported for both built-in and external databases. This issue will be addressed in a future release.
- LLM Streaming Response Payload Received All at Once: We have identified an issue where streaming responses are being received all at once, instead of as a continuous stream. See the No Streamed Responses Troubleshooting article for more information and a workaround.
- UI Mislabels Deployment on Errors Tab: While a deployment is in progress, the UI may display error messages for a “Deployment B” when viewing the Errors tab; this is a mislabeling of the deployment. In this case, you can ignore the initial column and instead focus on the “Type” and “Message” columns during troubleshooting.
- Custom Grafana Base URLs: Deployments come with pre-built Grafana dashboard quicklinks — however, if you choose to manually configure the
grafana.deployment_dashboard_baseurl
invalues.yaml
, the link cannot be set to point to a completely different Grafana instance with a full host path; it must point to a different path on the same Grafana instance.