Deploy machine learning models in production

Cortex is an open-source machine learning deployment platform you can self-host on AWS

Get Started GitHub

How it works

Define your deployment using declarative configuration

# cortex.yaml - kind: api name: my-api model: s3://my-bucket/my-model.onnx request_handler: handler.py

Customize request handling before and after inference

# handler.py def pre_inference(sample, metadata): # Python code def post_inference(prediction, metadata): # Python code

Deploy to your AWS account

$ cortex deploy Deploying ... Ready! https://***.amazonaws.com/my-api

Serve real time predictions via scalable JSON APIs

$ curl -d '{"a": 1, "b": 2, "c": 3}' \ https://***.amazonaws.com/my-api { prediction: "def" }

Supported frameworks

Key features

Declarative

Deployments are defined using declarative configuration

Autoscaling

Cortex automatically scales APIs for production workloads

Rolling updates

APIs update without any downtime

Cloud native

Cortex can be deployed on any AWS account in minutes