Model Serving Guide#
Assuming you have already created a model in the Model Registry, a deployment can now be created to prepare a model artifact for this model and make it accessible for running predictions behind a REST endpoint. Follow the Deployment Creation Guide to create a Deployment for your model.
Predictors are responsible for running a model server that loads a trained model, handles inference requests and returns predictions, see the Predictor Guide.
Transformers are used to apply transformations on the model inputs before sending them to the predictor for making predictions using the model, see the Transformer Guide.
Configure the predictor to batch inference requests, see the Inference Batcher Guide.
Configure the predictor to log inference requests and predictions, see the Inference Logger Guide.