Use a deployment for on-demand inference
After you deploy your custom model for on-demand inference, you can use it to generate
responses by making inference requests. For InvokeModel or
Converse operations, you use the deployment Amazon Resource Name (ARN)
as the modelId.
For information about making inference requests, see the following topics: