Deploy a Model in Studio - Amazon SageMaker AI

Deploy a Model in Studio

After you register a model version and approve it for deployment, deploy it to a Amazon SageMaker AI endpoint for real-time inference. You can Deploy a Model from the Registry with Python or deploy your model in Amazon SageMaker Studio. The following provides instructions on how to deploy your model in Studio.

This feature is not available in Amazon SageMaker Studio Classic.

Before you can deploy a model package, the following requirements must be met for the model package:

The following provides instructions on how to deploy a model in Studio.

To deploy a model in Studio
  1. Open the Studio console by following the instructions in Launch Amazon SageMaker Studio.

  2. Choose Models from the left navigation pane.

  3. Choose the Registered models tab, if not selected already.

  4. Immediately below the Registered models tab label, choose Model Groups, if not selected already.

  5. (Optional) If you have models that are shared with you, you can choose between My models or Shared with me.

  6. Select the checkboxes for the registered models. If the above requirements are met, the Deploy button becomes available to choose.

  7. Choose Deploy to open the Deploy model to endpoint page.

  8. Configure the deployment resources in the Endpoint settings.

  9. Once you have verified the settings, choose Deploy. The model will then be deployed to the endpoint with the In service status.

For us-east-1, us-west-2, ap-northeast-1, and eu-west-1 regions, you can use the following instructions to deploy models:

To deploy a model in Studio
  1. Open the Studio console by following the instructions in Launch Amazon SageMaker Studio.

  2. Choose Models from the left navigation pane.

  3. Choose the My models tab.

  4. Choose the Logged models tab, if not selected already.

  5. Select a model and choose View Latest Version.

  6. Choose Deploy and select between SageMaker AI or Amazon Bedrock.

  7. Once you have verified the settings, choose Deploy. The model will then be deployed to the endpoint with the In service status.