This document discusses deploying deep learning models in ONNX format with the Apache MXNet Model Server. It describes how ONNX serves as a common format that can be exported from many frameworks and imported into MXNet. It also outlines how the MXNet Model Server provides an API and containerization for serving models, and how it can be deployed in serverless environments on AWS Fargate for scalable inference.