Deploying AutoGluon models with serverless templates

Open In Colab Open In SageMaker Studio Lab

After learning how to train/deploy models using AWS SageMaker Cloud Training with Amazon SageMaker, in this section we will learn how to deploy trained models using AWS Lambda.

Reducing the model size to minimize AWS Lambda startup times

When the Lambda service receives a request to run a function via the Lambda API, the service first prepares an execution environment. During this step, the service downloads the code for the function, which is stored in Amazon Elastic Container Registry. It then creates an environment with the memory, runtime, and configuration specified. Once complete, Lambda runs any initialization code outside of the event handler before finally running the handler code. The steps of setting up the environment and the code are frequently referred to as a “cold start”.

After the execution completes, the execution environment is frozen. To improve resource management and performance, the Lambda service retains the execution environment for a non-deterministic period of time. During this time, if another request arrives for the same function, the service may reuse the environment. This second request typically finishes more quickly, since the execution environment already exists and it’s not necessary to download the code and run the initialization code. This is called a “warm start”.

Because AutoGluon containers are larger than a typical Lambda container, it might take some time (60+ seconds) to perform steps required for a “cold start”.
This could be limiting factor when used with latency-sensitive applications. To reduce start up times with AWS Lambda it is important to reduce model size to a minimum. This can be done by applying deployment-optimized presets as described in section “Faster presets or hyperparameters” of Predicting Columns in a Table - In Depth:

presets = ['good_quality_faster_inference_only_refit', 'optimize_for_deployment']

If the cold boot latency cannot be tolerated, it is recommended to reserve concurrent capacity as described in this article: Managing Lambda reserved concurrency.

More details on the lambda performance optimizations can be found in the following article: Operating Lambda: Performance optimization

Creating a base project

To start the project, please follow the setup steps of the tutorial: Deploying machine learning models with serverless templates.

To deploy AutoGluon, the following adjustments would be required:

  • the trained model is expected to be in ag_models directory.

  • Dockerfile to package AutoGluon runtimes and model files

  • Modify serving app/app.py script to use AutoGluon

When building a docker container it’s size can be reduced using the following optimizations:

  • use CPU versions of pytorch; if the models to be deployed don’t use pytorch, then don’t install it.

  • install only the AutoGluon sub-modules required for inference - specifically autogluon.tabular[all] will deploy only all tabular models without text and vision modules and their extra dependencies. This instruction can be further narrowed down to a combination of the following options are: lightgbm, catboost, xgboost, fastai and skex.

The following Dockerfile can be used as a starting point:

FROM public.ecr.aws/lambda/python:3.8

RUN yum install libgomp git -y \
 && yum clean all -y && rm -rf /var/cache/yum

ARG TORCH_VER=1.9.1+cpu
ARG TORCH_VISION_VER=0.10.1+cpu
ARG NUMPY_VER=1.19.5
RUN python3.8 -m pip --no-cache-dir install --upgrade --trusted-host pypi.org --trusted-host files.pythonhosted.org pip \
 && python3.8 -m pip --no-cache-dir install --upgrade wheel setuptools \
 && python3.8 -m pip uninstall -y dataclasses \
 && python3.8 -m pip --no-cache-dir install --upgrade torch=="${TORCH_VER}" torchvision=="${TORCH_VISION_VER}" -f https://download.pytorch.org/whl/torch_stable.html \
 && python3.8 -m pip --no-cache-dir install --upgrade numpy==${NUMPY_VER} \
 && python3.8 -m pip --no-cache-dir install --upgrade autogluon.tabular[all]"

COPY app.py ./
COPY ag_models /opt/ml/model/

CMD ["app.lambda_handler"]

Lambda serving script (app/app.py):

import pandas as pd
from autogluon.tabular import TabularPredictor

model = TabularPredictor.load('/opt/ml/model')
model.persist(models='all')


# Lambda handler code
def lambda_handler(event, context):
    df = pd.read_json(event['body'])
    pred_probs = model.predict_proba(df)
    return {
        'statusCode': 200,
        'body': pred_probs.to_json()
    }

Once the necessary modifications to the projects are done, proceed with the steps described in “Deploying the application to Lambda” section of the tutorial.

Conclusion

In this tutorial we explored how to deploy AutoGluon models as a serverless application. To explore more, refer to the following documentation: