Build, Serve and Deploy

Introduction to Ai-APIᵀᴹ

Overview

AI-API™ makes moving trained ML models to production easy:

  • Package models trained with an ML framework and then containerize the model server for production deployment

  • Deploy anywhere for online API serving endpoints or offline batch inference jobs

  • High-Performance API model server with adaptive micro-batching support

  • AI-API™ server can handle high volume without crashing, supports multi-model inference, API server Dockerization, Built-in Prometheus metric endpoint, Swagger/Open API endpoint for API Client library generation, serverless endpoint deployment, etc.

  • Central hub for managing models and the deployment process via web UI and APIs

  • Supports various ML frameworks, including:

Scikit-Learn, PyTorch, TensorFlow 2.0, Keras, FastAI v1 & v2, XGBoost, H2O, ONNX, Gluon and more

  • Supports API input data types, including:

DataframeInput, JsonInput, TfTensorflowInput, ImageInput, FileInput, MultifileInput, StringInput, AnnotatedImageInput, and more

  • Supports API output adapters, including:

BaseOutputAdapter, DefaultOutput, DataframeOutput, TfTensorOutput, and JsonOutput

Easy steps to AI-API Deployment

Last updated

Was this helpful?