# Introduction

### Overview

Ai-API™ makes moving trained ML models to production easy:

* Package models trained with ML framework and then containerize the model server for production deployment&#x20;
* Deploy anywhere for online API serving endpoints or offline batch inference jobs
* High-Performance API model server with adaptive micro-batching support
* Ai-API™ server is able to handle high-volume without crashing, supports multi-model inference, API server Dockerization, Built-in Prometheus metric endpoint, Swagger/Open API endpoint for API Client library generation, serverless endpoint deployment etc.
* Central hub for managing models and deployment process via web UI and APIs
* Supports various ML frameworks including:

Scikit-Learn, PyTorch, TensorFlow 2.0, Keras, FastAI v1 & v2, XGBoost, H2O, ONNX, Gluon and more

* Supports API input data types including:&#x20;

DataframeInput, JsonInput, TfTensorflowInput, ImageInput, FileInput, MultifileInput, StringInput, AnnotatedImageInput and more

* Supports API output Adapters including:&#x20;

BaseOutputAdapter, DefaultOutput, DataframeOutput, TfTensorOutput and JsonOutput

### Easy steps to Ai-AP&#x49;**™** Deployment

1. [Select your notebook](https://computationaldocs.zeblok.com/info/1.3.0/ai-api-engine/notebooks)
2. [Build Model](https://computationaldocs.zeblok.com/info/1.3.0/ai-api-engine/build-models)
3. [Deploy](https://computationaldocs.zeblok.com/info/1.3.0/ai-api-engine/deploy)
