How to train ML model

In this article, you will learn how to manage ML model training with Quix. In this example, we will train a model to predict car braking on a racing circuit 5 seconds ahead of time.

Why this is important

With the Quix platform, you can leverage historic data to train your model to react to data coming from source with milliseconds latency.

End result

At the end of this article, we will end up with a pickle file trained on historic data.

model training


You will need Python3 installed.

You’ll need some data stored in the Quix platform. You can use our Quick Start guide to do this using only the Quix portal.

You’ll also need a Jupyter notebook environment to run your experiments and load data for training. Please use "How to work with Jupyter notebook".

Install required libraries

python3 -m pip install seaborn
python3 -m pip install sklearn
python3 -m pip install mlflow
python3 -m pip install matplotlib

If you don’t see Python3 kernel in your Jupyter notebook, execute the following commands in your python environment:

python3 -m pip install ipykernel
python3 -m ipykernel install --user

Necessary imports

To execute all code blocks below, you need to start with importing these libraries. Add this code to the top of you Jupyter notebook.

import math
import matplotlib.pyplot as plt
import mlflow
import numpy as np
import pandas as pd
import pickle
import seaborn as sns

from sklearn import tree
from sklearn.model_selection import KFold
from sklearn.metrics import confusion_matrix, accuracy_score
from sklearn.tree import DecisionTreeClassifier

Training ML model

Getting training data

Quix web application has a python code generator to help you connect your Jupyter notebook to Quix.

To generate python code, follow those steps:
  1. Go to the platform

  2. Select workspace

  3. Go to the Data catalogue data icon

  4. Select car data stream to visualize visualize icon

  5. Select Brake, Motion_Yaw, Steer, Speed, Gear parameters

  6. Press Connect button

  7. Select Python language

  8. Copy Python code to your Jupyter notebook and execute.

connect python
jupyter results
If you want to use this generated code for a long time, replace the temporary token with a PAT token. See authenticate your requests for how to do that.

Preprocessing of features

We will prepare data for training by applying some transformation on downloaded data.

# Convert yaw angle to continuous values
df["Motion_Yaw_sin"] = df["Motion_Yaw"].map(lambda x: math.sin(x))
df["Motion_Yaw_cos"] = df["Motion_Yaw"].map(lambda x: math.cos(x))

Preprocessing of label

Here we simplify breaking to boolean.

# Conversion of label
df["Brake_bool"] = df["Brake"].map(lambda x: round(x))

Generate advanced brake signal for training

Now we need to shift breaking 5 seconds ahead to train model to predict breaking 5 seconds ahead.

# Offset dataset and trim it
NUM_PERIODS = -round(5e9/53852065.77281786)

df["Brake_shifted_5s"] = df["Brake_bool"].shift(periods=NUM_PERIODS)
df = df.dropna(axis='rows')

Lets review it in plot:

plt.figure(figsize=(15, 8))
plt.legend(['Shifted', 'Unshifted'])
brake shifted

Fit, predict and score a model

Calculate class weighting in case we gain any accuracy by performing class balancing.

Y = df["Brake_shifted_5s"]

cw = {}
for val in set(Y):
    cw[val] = np.sum(Y != val)



In the following code snippet we are executing an experiment using MLflow. Notice in last 3 lines that each experiment is logging MLflow metrics for experiments comparison later.

model_accuracy = pd.DataFrame(columns=[
    'Baseline Training Accuracy',
    'Model Training Accuracy',
    'Baseline Testing Accuracy',
    'Model Testing Accuracy',

kfold = KFold(5, shuffle=True, random_state=1)

with mlflow.start_run():
    class_weight = None
    max_depth = 5
    features = ["Motion_Yaw_cos", "Motion_Yaw_sin", "Steer", "Speed", "Gear"]

    mlflow.log_param("class_weight", class_weight)
    mlflow.log_param("max_depth", max_depth)
    mlflow.log_param("features", features)
    mlflow.log_param("model_type", "DecisionTreeClassifier")

    X = df[features]
    decision_tree = DecisionTreeClassifier(class_weight=class_weight, max_depth=max_depth)

    for train, test in kfold.split(X):
        X_train = X.iloc[train]
        Y_train = Y.iloc[train]
        X_test = X.iloc[test]
        Y_test = Y.iloc[test]

        # Train model, Y_train)
        Y_pred = decision_tree.predict(X_test)

        # Assess accuracy
        train_accuracy = round(decision_tree.score(X_train, Y_train) * 100, 2)
        test_accuracy = round(decision_tree.score(X_test, Y_test) * 100, 2)

        Y_baseline_zeros = np.zeros(Y_train.shape)
        baseline_train_accuracy = round(accuracy_score(Y_train, Y_baseline_zeros) * 100, 2)
        Y_baseline_zeros = np.zeros(Y_test.shape)
        baseline_test_accuracy = round(accuracy_score(Y_test, Y_baseline_zeros) * 100, 2)

        model_accuracy = model_accuracy.append({
            "Baseline Training Accuracy": baseline_train_accuracy,
            "Model Training Accuracy": train_accuracy,
            "Baseline Testing Accuracy": baseline_test_accuracy,
            "Model Testing Accuracy": test_accuracy
        }, ignore_index=True)

    mlflow.log_metric("train_accuracy", model_accuracy["Model Training Accuracy"].mean())
    mlflow.log_metric("test_accuracy", model_accuracy["Model Testing Accuracy"].mean())
    mlflow.log_metric("fit_quality", 1/abs(model_accuracy["Model Training Accuracy"].mean() - model_accuracy["Model Testing Accuracy"].mean()))

We review experiment model accuracy:

Table 1. Table with model accuracy preview
Depth Baseline Training Accuracy Model Training Accuracy Baseline Testing Accuracy Model Testing Accuracy


























Prediction preview

Let’s plot actual vs predicted braking using a trained model:

f, (ax1, ax2) = plt.subplots(2, 1, sharey=True, figsize=(50,8))


Saving model

When you are confident with the results, save the model into a file and commit it to GIT.

pickle.dump(decision_tree, open('./decision_tree_5_depth.sav', 'wb'))
Pickle file will be located in folder where jupyter notebook command was executed


To help you with experiments management, you can review experiments in MLflow.

MLflow is working only in MacOS, Linux or Windows linux subsystem.
To have some meaningful data, run the experiment with 3 different max_depth parameter.

Let’s leave Jupyter notebook for now and go back to command line and run MLflow server:

mlflow ui

Select experiments to compare:


Plot metrics from experiments:

experiments comparison