Deploy a custom model for on-demand inference
After you create a custom model with a model customization job or import a SageMaker AI-trained custom Amazon Nova model, you can set up on-demand inference for the model. With on-demand inference, you only pay for what you use and you don't need to set up provisioned compute resources.
To set up on-demand inference for a custom model, you deploy it with a custom model deployment. After
you deploy your custom model, you use the deployment's Amazon Resource Name (ARN) as the modelId
parameter when you submit prompts and generate responses with model inference.
For information about on-demand inference pricing, see Amazon Bedrock pricing
-
US East (N. Virginia)
Topics
Supported base models
You can set up on-demand inference for the following base models:
-
Amazon Nova Lite
-
Amazon Nova Micro
-
Amazon Nova Pro