Deploy a custom model for on-demand inference
After you successfully create a custom model with a model customization job (fine-tuning, distillation, or continued pre-training), you can set up on-demand inference for the model.
To set up on-demand inference for a custom model, you deploy the model with a custom model
deployment. After you deploy your custom model, you use the deployment's Amazon Resource
Name (ARN) as the modelId
parameter in your InvokeModel
or
Converse
API operations. You can use the deployed model for on-demand
inference with Amazon Bedrock features such as playgrounds, Agents, and Knowledge Bases.
Topics
Supported models
You can set up on-demand inference for the following models:
-
Amazon Nova Canvas
-
Amazon Nova Lite
-
Amazon Nova Micro
-
Amazon Nova Pro