Deploy
                        publicly available foundation models with the JumpStartModel
                        class
You can deploy a built-in algorithm or pre-trained model to a SageMaker AI endpoint in just a few lines of code using the SageMaker Python SDK.
- 
                        First, find the model ID for the model of your choice in the Built-in Algorithms with pre-trained Model Table . 
- 
                        Using the model ID, define your model as a JumpStart model. from sagemaker.jumpstart.model import JumpStartModel model_id ="huggingface-text2text-flan-t5-xl"my_model = JumpStartModel(model_id=model_id)
- 
                        Use the deploymethod to automatically deploy your model for inference. In this example, we use the FLAN-T5 XL model from Hugging Face.predictor = my_model.deploy()
- 
                        You can then run inference with the deployed model using the predictmethod.question ="What is Southern California often abbreviated as?"response = predictor.predict(question) print(response)
Note
This example uses the foundation model FLAN-T5 XL, which is suitable for a wide range of text generation use cases including question answering, summarization, chatbot creation, and more. For more information about model use cases, see Available foundation models.
For more information about the JumpStartModel class and its
                    parameters, see JumpStartModel
Check default instance types
You can optionally include specific model versions or instance types when
                        deploying a pre-trained model using the JumpStartModel class.
                        All JumpStart models have a default instance type. Retrieve the default
                        deployment instance type using the following code:
from sagemaker import instance_types instance_type = instance_types.retrieve_default( model_id=model_id, model_version=model_version, scope="inference") print(instance_type)
See all supported instance types for a given JumpStart model with the
                            instance_types.retrieve() method.
Use inference components to deploy multiple models to a shared endpoint
An inference component is a SageMaker AI hosting object that you can use to deploy
                        one or more models to an endpoint for increased flexibility and scalability.
                        You must change the endpoint_type for your JumpStart model to be
                        inference-component-based rather than the default model-based endpoint. 
predictor = my_model.deploy( endpoint_name ='jumpstart-model-id-123456789012', endpoint_type =EndpointType.INFERENCE_COMPONENT_BASED)
For more information on creating endpoints with inference components and deploying SageMaker AI models, see Shared resource utilization with multiple models.
Check valid input and output inference formats
To check valid data input and output formats for inference, you can use
                        the retrieve_options() method from the Serializers
                        and Deserializers classes.
print(sagemaker.serializers.retrieve_options(model_id=model_id, model_version=model_version)) print(sagemaker.deserializers.retrieve_options(model_id=model_id, model_version=model_version))
Check supported content and accept types
Similarly, you can use the retrieve_options() method to check
                        the supported content and accept types for a model.
print(sagemaker.content_types.retrieve_options(model_id=model_id, model_version=model_version)) print(sagemaker.accept_types.retrieve_options(model_id=model_id, model_version=model_version))
For more information about utilities, see Utility APIs