DEV Community

Sabah Shariq
Sabah Shariq

Posted on

Training a Machine Learning model in Azure

Introduction

We will use a credit card dataset to understand how to use Machine Learning for a classification problem in Azure. Here our goal is to identify customers who are at risk of not paying their credit card bills. To get started with Azure Machine Learning following two prerequisites first we need to setup:

Workspace

In order to use Machine Learning in Azure we need to create a workspace. A workspace is a centralized environment where all the machine learning resource get stored.

Compute Instance

A compute instance is a cloud-based virtual machine (VM) used for developing and running ML models. Here we will run Jupyter notebooks and Python scripts and test AI algorithms.

Create a workspace

  • Sign in to Azure Machine Learning studio
  • Select Create workspace
  • Provide the following information to configure your new workspace:
  • Workspace name
  • Friendly name
  • Hub
    • If you did not select a hub, provide the advanced information
    • If you selected a hub, these values are taken from the hub.
  • Subscription
  • Resource group
  • Region
  • Select Create to create the workspace

Create a compute instance

Create compute instance following steps:

  • Select your workspace
  • On the top right, select New
  • Select Compute instance in the list
  • Supply a name
  • Keep the default values for the rest of the page, unless your organization policy requires you to change other settings
  • Select Review + Create
  • Select Create

Create handle to workspace

We need a way to reference our workspace. That is Azure Machine Learning workspace is where all your ML resources live (models, datasets, compute, etc.). In order to work in the workspace "Python code" needs to connect to it and creating an "ml_client" does this.

from azure.ai.ml import MLClient
from azure.identity import DefaultAzureCredential

# authenticate
credential = DefaultAzureCredential()

SUBSCRIPTION="<SUBSCRIPTION_ID>"
RESOURCE_GROUP="<RESOURCE_GROUP>"
WS_NAME="<AML_WORKSPACE_NAME>"
# Get a handle to the workspace
ml_client = MLClient(
    credential=credential,
    subscription_id=SUBSCRIPTION,
    resource_group_name=RESOURCE_GROUP,
    workspace_name=WS_NAME,
)
Enter fullscreen mode Exit fullscreen mode

Here, we need to import required library

  • Import Libraries
  • MLClient → Used to interact with Azure Machine Learning.
  • DefaultAzureCredential → Handles authentication using Azure credentials
  • Authenticate
  • This automatically retrieves credentials from the user's Azure environment.
  • Define Azure Subscription & Resource Details
  • Connect to the Azure ML Workspace [ChatGPT explain]
  • Verify Connection [move code with above]
ws = ml_client.workspaces.get(WS_NAME)
print(ws.location,":", ws.resource_group)
Enter fullscreen mode Exit fullscreen mode
  • ml_client.workspaces.get(WS_NAME) → Fetches the workspace details from Azure.
  • ws.location → Prints the region where the workspace is deployed (e.g., "eastus").
  • ws.resource_group → Prints the resource group name.

If this prints the correct location and resource group, the connection is successful!

Enviroment Setup

To run machine learning jobs in Azure, you need to define an environment. To make sure the job runs correctly, you need to tell Azure what software and libraries are required.

You create an environment in Azure ML to tell Azure what tools, packages, and Python version your code needs to run correctly and reliably.

In this example, you create a custom conda environment for your jobs, using a conda yaml file.

Reference this yaml file to create and register this custom environment in your workspace:

import os

dependencies_dir = "./dependencies"
os.makedirs(dependencies_dir, exist_ok=True)

from azure.ai.ml.entities import Environment

custom_env_name = "aml-scikit-learn"

custom_job_env = Environment(
    name=custom_env_name,
    description="Custom environment for Credit Card Defaults job",
    tags={"scikit-learn": "1.0.2"},
    conda_file=os.path.join(dependencies_dir, "conda.yaml"),
    image="mcr.microsoft.com/azureml/openmpi4.1.0-ubuntu20.04:latest",
)
custom_job_env = ml_client.environments.create_or_update(custom_job_env)
Enter fullscreen mode Exit fullscreen mode

This registers the environment to your Azure ML workspace using ml_client.

print(
    f"Environment with name {custom_job_env.name} is registered to workspace, the environment version is {custom_job_env.version}"
)
Enter fullscreen mode Exit fullscreen mode

Confirms that the environment has been successfully created or updated.

Configure a command job to run training script

You're setting up a command job in Azure to run a Python script that trains a credit default prediction model using GradientBoostingClassifier.
This training script

  • Prepares the data
  • Trains the model
  • Splits the data using train_test_split (into training and testing sets)
  • Registers the trained model so it can be reused or deployed later

Here we will use Azure ML Studio to create and run the command job.

%%writefile {train_src_dir}/main.py
import os
import argparse
import pandas as pd
import mlflow
import mlflow.sklearn
from sklearn.ensemble import GradientBoostingClassifier
from sklearn.metrics import classification_report
from sklearn.model_selection import train_test_split

def main():
    """Main function of the script."""

    # input and output arguments
    parser = argparse.ArgumentParser()
    parser.add_argument("--data", type=str, help="path to input data")
    parser.add_argument("--test_train_ratio", type=float, required=False, default=0.25)
    parser.add_argument("--n_estimators", required=False, default=100, type=int)
    parser.add_argument("--learning_rate", required=False, default=0.1, type=float)
    parser.add_argument("--registered_model_name", type=str, help="model name")
    args = parser.parse_args()

    # Start Logging
    mlflow.start_run()

    # enable autologging
    mlflow.sklearn.autolog()

    ###################
    #<prepare the data>
    ###################
    print(" ".join(f"{k}={v}" for k, v in vars(args).items()))

    print("input data:", args.data)

    credit_df = pd.read_csv(args.data, header=1, index_col=0)

    mlflow.log_metric("num_samples", credit_df.shape[0])
    mlflow.log_metric("num_features", credit_df.shape[1] - 1)

    #Split train and test datasets
    train_df, test_df = train_test_split(
        credit_df,
        test_size=args.test_train_ratio,
    )
    ####################
    #</prepare the data>
    ####################

    ##################
    #<train the model>
    ##################
    # Extracting the label column
    y_train = train_df.pop("default payment next month")

    # convert the dataframe values to array
    X_train = train_df.values

    # Extracting the label column
    y_test = test_df.pop("default payment next month")

    # convert the dataframe values to array
    X_test = test_df.values

    print(f"Training with data of shape {X_train.shape}")

    clf = GradientBoostingClassifier(
        n_estimators=args.n_estimators, learning_rate=args.learning_rate
    )
    clf.fit(X_train, y_train)

    y_pred = clf.predict(X_test)

    print(classification_report(y_test, y_pred))
    ###################
    #</train the model>
    ###################

    ##########################
    #<save and register model>
    ##########################
    # Registering the model to the workspace
    print("Registering the model via MLFlow")
    mlflow.sklearn.log_model(
        sk_model=clf,
        registered_model_name=args.registered_model_name,
        artifact_path=args.registered_model_name,
    )

    # Saving the model to a file
    mlflow.sklearn.save_model(
        sk_model=clf,
        path=os.path.join(args.registered_model_name, "trained_model"),
    )
    ###########################
    #</save and register model>
    ###########################

    # Stop Logging
    mlflow.end_run()

if __name__ == "__main__":
    main()
Enter fullscreen mode Exit fullscreen mode

Above code explanation

Set Up the Script
Uses argparse to get inputs (like data path, model name, and training settings).

Starts an MLflow logging session.

Prepare the Data

credit_df = pd.read_csv(args.data)
train_df, test_df = train_test_split(credit_df)
Reads the CSV file.
Enter fullscreen mode Exit fullscreen mode

Splits the data into training and test sets using train_test_split.

Train the Model

clf = GradientBoostingClassifier()
clf.fit(X_train, y_train)
Enter fullscreen mode Exit fullscreen mode

Trains a Gradient Boosting Classifier, a tree-based ML model.

Then uses the model to make predictions.

Log Metrics and Results with MLflow

mlflow.sklearn.autolog()
mlflow.log_metric("num_samples", ...)
Enter fullscreen mode Exit fullscreen mode

Automatically logs model metrics like accuracy, training time, etc.

Helps track and compare different model runs.

Register the Trained Model

mlflow.sklearn.log_model(...)
mlflow.sklearn.save_model(...)
Enter fullscreen mode Exit fullscreen mode

The model is saved and registered to your Azure ML Workspace. This allows you to keep track of your trained models.

Configure a training job using "Command"

Now we need to configure a training job in Azure Machine Learning using the command function which will run the main.py training script in Azure with the right data, parameters and environment.

Create a variable called "inputs" that will use for:

  • data: The CSV file with credit card data.
  • test_train_ratio: How much of the data should be used for testing (20%).
  • learning_rate: A parameter for the model training.
  • registered_model_name: The name to register your trained model under.

Submit the job

Submit the job to run in Azure Machine Learning studio. This time, willuse create_or_update on ml_client. ml_client is a client class that allows you to connect to your Azure subscription using Python and interact with Azure Machine Learning services. ml_client allows to submit the jobs using Python.

ml_client.create_or_update(job)
Enter fullscreen mode Exit fullscreen mode

Conclusion

After run the cell, the notebook output shows a link to the job's details page on Machine Learning studio.

Heroku

Deploy with ease. Manage efficiently. Scale faster.

Leave the infrastructure headaches to us, while you focus on pushing boundaries, realizing your vision, and making a lasting impression on your users.

Get Started

Top comments (0)

👋 Kindness is contagious

Please leave a ❤️ or a friendly comment on this post if you found it helpful!

Okay