cancel
Showing results for 
Search instead for 
Did you mean: 
dylanwan
Employee
Employee

Introduction

This article discusses how you can use Incorta's built-in integration with MLflow to manage your machine learning lifecycle.

What is MLflow?

MLflow is an open-source platform that helps you keep track of your machine learning experiments by logging parameters, metrics, and even the resulting model artifacts (like pickled files or model definitions). This allows you to easily compare different runs, understand what led to better performance, and reproduce past results. The platform also provides a central model registry where you can store, version, and manage your trained models, bringing structure and organization to the often chaotic process of developing and deploying machine learning models.

Architecture

Incorta ML model development is done via Incorta Materialized View or in Incorta notebook for business users. The ML model artifacts are stored in your Incorta tenant storage. Loading the ML model in the MVs can integrate ML inference with your regular data pipeline.

You can definitely perform ML model development and inference with or without enabling MLflow. However, by enabling the MLflow integration option from the Incorta Cloud admin console, Incorta runs an ML tracking server for you.

Your MV will interact with the MLflow tracking server via the MLflow API, which is a standard API from the open-source MLflow package.

Screenshot 2025-04-29 at 3.24.10 PM.png

MLflow tracking includes a metadata database for holding ML models and cloud storage for storing MLflow artifacts.

Screenshot 2025-04-29 at 3.31.13 PM.png

The MLflow UI, which allows you to view the stored ML model and compare its performance, will be available. 

Screenshot 2025-04-29 at 5.16.24 PM.png

Screenshot 2025-04-29 at 5.22.27 PM.png

Configuration

Enable MLFlow

You can enable MLflow from the Incorta cloud admin console.  

Screenshot 2025-04-29 at 4.26.34 PM.png

You will be able to access the MLflow UI page. The login information will be emailed to the admin user. 

The tracking server address will be available to you from the page.  

Screenshot 2025-04-29 at 5.07.01 PM.png

 

MLflow and Chidori

Incorta Cloud provides the option of running your Spark job driver process on a separate machine hosted in the Incorta Cloud. When the MLflow option is enabled, the Chidori mode will be enabled. Your Spark driver processes will no longer run on the analytics service or loader service nodes. The change is transparent to the Incorta users, as  Incorta manages the complexity for you.

 

Related Material

Getting Started with Machine Learning in Incorta Guide

Action on Insights: Incorta for Data Scientists

Build Machine Learning Models using Incorta Materialized Views

Save Spark ML Model

Incorta third-party AI/ML platform Integration

 

 

Best Practices Index
Best Practices

Just here to browse knowledge? This might help!

Contributors
Version history
Last update:
‎05-07-2025 05:57 AM
Updated by: