Tech Blog: Deploying an AWS Lambda function in Python for a machine learning model
Posted by Abhishek R. S. on 2024-07-04, updated on 2024-07-22

The blog is about...

Creating a serverless deployment using AWS Lamdba function in Python for serving a machine learning model. The project can be found in the following Github repo ML_NY_TAXI/aws_lambda_func .

1.1) Creating a lambda function script in Python

A lambda function can be created in Python very easily. Refer to the project repo. A function named lambda_handler which takes an event as the argument returns a prediction inside a dict. To perform sanity test for the lambda function locally, run the script named sanity_lambda_func.py. Once the sanity check for the working of the lambda function script is successful, it is ready to be containerized with docker.

1.2) IAM setup

Although, this is out of the scope of this blogpost, it is a standard practice to setup IAM users, groups and roles. Create an AWS user if not present already. Create a group for AWS Lambda users, API Gateway admin, ECR users. Add user to these groups. Add policy to these groups by giving full access to relevant things. Create a role for lambda with full access to lambda.

1.3) Creating a docker image for the lambda function and testing it locally

A list of publicly available docker images for AWS can be found in Amazon Elastic Container Registry (ECR) public gallery. One can search for "lambda python" and find a docker image for appropriate python version under Image tags. This image can be used as a base docker image to build a custom docker image for a lambda function.

Unlike a typical Dockerfile, there will be certain changes.

  • There is no need to set working directory using WORKDIR /app that is usually done in case of dockerizing a Flask / FastAPI application. There is a default working directory that is configured in the base docker image.

  • There is no need to expose a port for dockerizing the lambda function, that is usually done in case of dockerizing a Flask / FastAPI application.

  • The final CMD to run an application inside the container will just be the
    lambda_function_script:lambda_function_handler

To build the docker container, use the following command,
docker build -t aws-lambda-ny-taxi .

To send post request to the lambda function running inside the docker container locally, run the following command,
docker run -it --rm -p 8080:8080 aws-lambda-ny-taxi:latest
followed by the post request script named post_req_lambda_func_local.py

1.4) Creating a repo for docker container on AWS ECR

Create a private repo for the docker container using the following steps. ECR -> Create repo -> Select private -> Choose a name for the repo -> Create repo

Once the repo is created, go to the repo on AWS, see View Push Commands. Run all the commands in the terminal.

1.5) Setting up the Lambda function on AWS

The AWS Lambda function needs to be set up. This can be done using the following steps. Lambda -> Create function -> Select from container image -> Choose a name for the lambda function -> Choose the container image with browse images -> Choose x86_64 -> Choose existing role and select appropriate role -> Create

One can easily test the lambda function using the test option. Copy paste the json input and then test. Doing this directly will result in a failure since it will timeout (default timeout is 3 sec.). Set a higher timeout like 30 sec. (for initial loading) and a slightly higher memory in Configuration -> General -> Timeout. After applying the above changes and test it, things will work. The first time will take a longer time to run (like > 20 sec.) and the subsequent runs will be much faster (in milli sec.). The following image shows the successful test of the lambda function.

1.6) Exposing the AWS Lambda function using AWS API Gateway

A Rest API needs to be created for exposing the Lambda function. This can be done using the following steps. API Gateway -> Rest API -> Choose a name for the API -> Create API

Once the API is created, a resource needs to be created. This can be done by create resource and choose the name of the resource (/predict).

Create a method to invoke the resource using the following steps. Create method -> Select method type (POST) -> Choose as Lambda function -> Select the appropriate lambda function -> Create method The API can be tested using Test. Copy paste the json to the request body and test. The first time will take a longer time to run (>2 sec.). The subsequent runs will take much less time (in milli sec.). The following image shows the successful test of the deployed API using the AWS API gateway.

After all these steps, one can deploy the API using Deploy API -> select new stage -> choose a stage name (test) Use the invoke URL in post_req_lambda_func_aws.py script to test sending POST requests to the API deployed on AWS.

1.7) Challenges faced

  • I forgot to remove the WORKDIR /app from the Dockerfile for the lambda function and got some errors that the script lambda_function.py was not found. I figured it out later and removed WORKDIR /app. This is because of the fact that AWS is providing the base docker image and they would have built it with their own config.

  • Another mistake that I made was that number of arguments for any lambda function lambda_handler() is two and it will require event and context as arguments. I initially used only one (just the event without context) as I was still learning and experimenting. I got some error regarding mismatch in the number of function args. After fixing this, things worked smoothly.

Main takeaway

  • I learned to create a lambda function in Python for deploying a machine learning model to AWS Lambda.

  • I learned to create a repo in ECR and push a docker container image to the ECR.

  • I learned to create an AWS Lambda function from the container image and test it.

  • I learned to create a Rest API for the deployed lambda function using the AWS API Gateway.

I learned a lot of new things in this project. To deploying more complex ML models using AWS Lambda in the future 😄.

Next Steps

  • Deploy multiple complex ML models to AWS Lambda.

  • Deploy lambda functions with different variants like adding a trigger to S3 bucket or streaming using Kinesis.

About the author: Abhishek is a graduate AI engineer who is interested in developing and deploying Artificial Intelligence, Machine Learning, Computer Vision, Data Science applications. His goal is to improve human lives by leveraging real world data.