MLRun v1.7 now available! Solidifying Generative AI Implementation and LLM Monitoring

Launching MLRun 1.7: Gen AI and LLM Monitoring

V1.7 brings significant LLM monitoring enhancements, helping users ensure the integrity and operational stability of LLMs in production environments.

As the open-source maintainers of MLRun, we’re proud to announce the release of MLRun v1.7.  MLRun is an open-source AI orchestration tool that accelerates the deployment of gen AI applications, with features such as LLM monitoring, data management, guardrails and more. We provide ready-made scenarios that can be easily implemented by teams in organizations. This new release is packed with powerful features designed to make gen AI deployments more flexible and faster than ever before.

Specifically, V1.7 brings significant LLM monitoring enhancements, helping users ensure the integrity and operational stability of LLMs in production environments. Additional updates introduce performance optimizations, multi-project management, and more.

Read all the details below:

1. Flexible Monitoring Infrastructure

MLRun 1.7 introduces a new, flexible monitoring infrastructure that enables seamless integration of external tools and applications into AI pipelines, using APIs and pre-built integration points. This includes tools for external logging, alerting, metrics systems, etc. 

For instance, users can now:

  • Track custom metrics that are specifically tailored to business needs, such as user-defined success metrics or domain-specific KPIs.
  • Integrate with open-source tools like Evidently, which enables advanced tracking of model performance metrics (e.g., distribution shifts, data quality, and accuracy).
  • Leverage external logging services to centralize logs and improve the visibility of pipeline activities

2. Better Monitoring of Unstructured Data

Given that LLMs primarily handle unstructured data, one of the key advances in MLRun 1.7 is its enhanced ability to enable tracking this kind of data with more precision.

A common way to monitor LLMs is to create another model that would act as a judge. See a demo of how this works.

3. Endpoint Metrics UI and Customization

MLRun 1.7 introduces a new endpoint metrics UI. Its expanded endpoint monitoring capabilities allow users to:

  • Select and investigate different endpoint metrics, such as accuracy and response times.
  • View various metrics related to model endpoints, such as the number of activations or event counts.
  • Visualize trends through time series and histogram views
  • Customize the monitoring time frame, such as looking at data from the past week or another specified period.

For example, a time-series chart could indicate a bottleneck in the inference pipeline or model scaling issues.

The ability to track, visualize, and analyze endpoint performance enables teams to adjust operational parameters or retrain models as soon as performance starts to degrade. This reduces downtime or adverse effects in production environments.

With these capabilities, users can now customize their monitoring stacks per their business and tech stack requirements. Future releases will continue to enhance these capabilities, with more features and integrations for monitoring. This will allow for even greater flexibility and user control. So please share your feedback, so we can extend them based on your needs.

Spotlight: Gen AI Banking Chatbot Demo

See a gen AI banking chatbot that uses MLRun’s new monitoring capabilities for fine-tuning, ensuring it only answers banking-related questions. This helps address the risks associated with gen AI, like hallucinations, inaccuracies, bias, harmful content, and more.

Watch the demo here.

5. Simplified Docker Deployment Workflow

Version 1.7 simplifies the process of deploying Docker images, making it easier for users to run applications and models. Previously, deploying applications or models via Docker required manual configuration, with open-source Nuclio, and integration steps. Now, users can simply provide a Docker image and deploy it with minimal setup.

This improvement opens up development workflow possibilities. For example, users can more easily integrate custom UIs or dashboards that can interact with deployed models, allowing for more advanced and customized monitoring capabilities.

6. Cross-Project View

For enterprises working on multiple projects across diverse teams, keeping track of workflows and active jobs can become overwhelming. MLRun 1.7 introduces a cross-project view that consolidates all activities across projects into a single, centralized dashboard.

The cross-project view provides real-time visibility into all active jobs, workflows, and ML models across different projects. Users can:

  • Monitor multiple projects to see which workflows and jobs are running, completed, or failed.
  • Identify issues in specific projects quick and more effectively

This is especially valuable for organizations with complex environments where multiple teams may be working on different but interrelated projects.

7. Community-Driven Innovations and Performance Enhancements

Finally, MLRun 1.7 introduces improvements based on the invaluable feedback from you, our community users. We listened to the requirements and are releasing features that provide value in areas the community cares about most. This version introduces improved UI responsiveness, more efficient handling of large datasets, and a host of usability fixes. We look forward to your continued feedback on this version and the upcoming ones as well.

Join the Conversation

We’re looking forward to hearing your feedback about MLRun 1.7 and your future needs for the upcoming versions. Join the community and share your insights and requirements.

Read the full changelog.

Explore MLRun 1.7.

Recent Blog Posts
Deploying Hugging Face LLM Models with MLRun
Alexandra Quinn
November 1, 2024