Skip to content

Latest commit

 

History

History
72 lines (56 loc) · 2.99 KB

README.md

File metadata and controls

72 lines (56 loc) · 2.99 KB

Serving PyTorch with Actix-Web

This repository gives an example of training a machine learning model using PyTorch, exporting that model, then serving inference over a RESTful API using Actix-Web in rust.

Implemented by request based on the original TensorFlow example.

For more information on the tools used, check out the following repositories:

Overview

The repository has 3 sections:

  • ./training – Contains the script which trains a neural network to recognize digits.
  • ./server – Contains the RESTful API webserver rust code.
  • ./client – Contains a sample script that demonstrates making a request to the server.

The training script will output a saved neural network model in compiled TorchScript format. The server then loads this into memory on startup. The server accepts a JSON payload at the /mnist endpoint with a single key "image" that is a base64 encoded image (PNG or JPG). This image is decoded, rescaled to the correct input dimensions, converted to grayscale, normalized (matching the training data normalization), and finally submitted to the model for inference. Predictions are returned with a "label" integer value and a "confidence" float between 0 and 1.

Setup

This example assumes you have rust installed and python 3.6+ setup. To install the needed python dependencies:

pip install -r requirements.txt

Training

The model used is a simple convolutional neural network trained on the MNIST dataset. The data is automatically downloaded using the torchvision library. To train the model:

python training/train.py

This will output a saved model to ./saved_model/model.pt. A pre-trained model is included in this repository. The model isn't too large and can be trained without any GPU.

Serving

The server code is a rust crate located in ./server. In order to run, the server requires the saved model directory location specified with the --model-path command line argument. you can try running the server with:

cd server
cargo run -- --model-dir ../saved_model/model.pt

Serving in Docker

For actual deployments, you probably would want to build a release in a container to serve the API. To build the docker image:

docker build -t actix-torch .

Then to run the image locally for testing:

docker run --rm -it -p 8080:8080 actix-torch

Client Testing

With the server running locally, you can test inference using ./client/client.py. Included is a PNG file with a handwritten "3" that is base64-encoded and submitted to the server.

To test:

python client/client.py

Input:

Expected output:

POST to http://127.0.0.1:8080/mnist
Response (200)
Content: {
    "label": 3,
    "confidence": 0.9999999
}