Show / Hide Table of Contents

Class CfnInferenceComponentPropsMixin.InferenceComponentSpecificationProperty

Details about the resources to deploy with this inference component, including the model, container, and compute resources.

Inheritance
object
CfnInferenceComponentPropsMixin.InferenceComponentSpecificationProperty
Implements
CfnInferenceComponentPropsMixin.IInferenceComponentSpecificationProperty
Inherited Members
object.GetType()
object.MemberwiseClone()
object.ToString()
object.Equals(object)
object.Equals(object, object)
object.ReferenceEquals(object, object)
object.GetHashCode()
Namespace: Amazon.CDK.CfnPropertyMixins.AWS.SageMaker
Assembly: Amazon.CDK.CfnPropertyMixins.dll
Syntax (csharp)
public class CfnInferenceComponentPropsMixin.InferenceComponentSpecificationProperty : CfnInferenceComponentPropsMixin.IInferenceComponentSpecificationProperty
Syntax (vb)
Public Class CfnInferenceComponentPropsMixin.InferenceComponentSpecificationProperty Implements CfnInferenceComponentPropsMixin.IInferenceComponentSpecificationProperty
Remarks

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html

ExampleMetadata: fixture=_generated

Examples
// The code below shows an example of how to instantiate this type.
             // The values are placeholders you should change.
             using Amazon.CDK.CfnPropertyMixins.AWS.SageMaker;

             var inferenceComponentSpecificationProperty = new InferenceComponentSpecificationProperty {
                 BaseInferenceComponentName = "baseInferenceComponentName",
                 ComputeResourceRequirements = new InferenceComponentComputeResourceRequirementsProperty {
                     MaxMemoryRequiredInMb = 123,
                     MinMemoryRequiredInMb = 123,
                     NumberOfAcceleratorDevicesRequired = 123,
                     NumberOfCpuCoresRequired = 123
                 },
                 Container = new InferenceComponentContainerSpecificationProperty {
                     ArtifactUrl = "artifactUrl",
                     DeployedImage = new DeployedImageProperty {
                         ResolutionTime = "resolutionTime",
                         ResolvedImage = "resolvedImage",
                         SpecifiedImage = "specifiedImage"
                     },
                     Environment = new Dictionary<string, string> {
                         { "environmentKey", "environment" }
                     },
                     Image = "image"
                 },
                 ModelName = "modelName",
                 StartupParameters = new InferenceComponentStartupParametersProperty {
                     ContainerStartupHealthCheckTimeoutInSeconds = 123,
                     ModelDataDownloadTimeoutInSeconds = 123
                 }
             };

Synopsis

Constructors

InferenceComponentSpecificationProperty()

Details about the resources to deploy with this inference component, including the model, container, and compute resources.

Properties

BaseInferenceComponentName

The name of an existing inference component that is to contain the inference component that you're creating with your request.

ComputeResourceRequirements

The compute resources allocated to run the model, plus any adapter models, that you assign to the inference component.

Container

Defines a container that provides the runtime environment for a model that you deploy with an inference component.

ModelName

The name of an existing SageMaker AI model object in your account that you want to deploy with the inference component.

StartupParameters

Settings that take effect while the model container starts up.

Constructors

InferenceComponentSpecificationProperty()

Details about the resources to deploy with this inference component, including the model, container, and compute resources.

public InferenceComponentSpecificationProperty()
Remarks

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html

ExampleMetadata: fixture=_generated

Examples
// The code below shows an example of how to instantiate this type.
             // The values are placeholders you should change.
             using Amazon.CDK.CfnPropertyMixins.AWS.SageMaker;

             var inferenceComponentSpecificationProperty = new InferenceComponentSpecificationProperty {
                 BaseInferenceComponentName = "baseInferenceComponentName",
                 ComputeResourceRequirements = new InferenceComponentComputeResourceRequirementsProperty {
                     MaxMemoryRequiredInMb = 123,
                     MinMemoryRequiredInMb = 123,
                     NumberOfAcceleratorDevicesRequired = 123,
                     NumberOfCpuCoresRequired = 123
                 },
                 Container = new InferenceComponentContainerSpecificationProperty {
                     ArtifactUrl = "artifactUrl",
                     DeployedImage = new DeployedImageProperty {
                         ResolutionTime = "resolutionTime",
                         ResolvedImage = "resolvedImage",
                         SpecifiedImage = "specifiedImage"
                     },
                     Environment = new Dictionary<string, string> {
                         { "environmentKey", "environment" }
                     },
                     Image = "image"
                 },
                 ModelName = "modelName",
                 StartupParameters = new InferenceComponentStartupParametersProperty {
                     ContainerStartupHealthCheckTimeoutInSeconds = 123,
                     ModelDataDownloadTimeoutInSeconds = 123
                 }
             };

Properties

BaseInferenceComponentName

The name of an existing inference component that is to contain the inference component that you're creating with your request.

public string? BaseInferenceComponentName { get; set; }
Property Value

string

Remarks

Specify this parameter only if your request is meant to create an adapter inference component. An adapter inference component contains the path to an adapter model. The purpose of the adapter model is to tailor the inference output of a base foundation model, which is hosted by the base inference component. The adapter inference component uses the compute resources that you assigned to the base inference component.

When you create an adapter inference component, use the Container parameter to specify the location of the adapter artifacts. In the parameter value, use the ArtifactUrl parameter of the InferenceComponentContainerSpecification data type.

Before you can create an adapter inference component, you must have an existing inference component that contains the foundation model that you want to adapt.

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html#cfn-sagemaker-inferencecomponent-inferencecomponentspecification-baseinferencecomponentname

ComputeResourceRequirements

The compute resources allocated to run the model, plus any adapter models, that you assign to the inference component.

public object? ComputeResourceRequirements { get; set; }
Property Value

object

Remarks

Omit this parameter if your request is meant to create an adapter inference component. An adapter inference component is loaded by a base inference component, and it uses the compute resources of the base inference component.

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html#cfn-sagemaker-inferencecomponent-inferencecomponentspecification-computeresourcerequirements

Type union: either IResolvable or CfnInferenceComponentPropsMixin.IInferenceComponentComputeResourceRequirementsProperty

Container

Defines a container that provides the runtime environment for a model that you deploy with an inference component.

public object? Container { get; set; }
Property Value

object

Remarks

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html#cfn-sagemaker-inferencecomponent-inferencecomponentspecification-container

Type union: either IResolvable or CfnInferenceComponentPropsMixin.IInferenceComponentContainerSpecificationProperty

ModelName

The name of an existing SageMaker AI model object in your account that you want to deploy with the inference component.

public string? ModelName { get; set; }
Property Value

string

Remarks

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html#cfn-sagemaker-inferencecomponent-inferencecomponentspecification-modelname

StartupParameters

Settings that take effect while the model container starts up.

public object? StartupParameters { get; set; }
Property Value

object

Remarks

See: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-sagemaker-inferencecomponent-inferencecomponentspecification.html#cfn-sagemaker-inferencecomponent-inferencecomponentspecification-startupparameters

Type union: either IResolvable or CfnInferenceComponentPropsMixin.IInferenceComponentStartupParametersProperty

Implements

CfnInferenceComponentPropsMixin.IInferenceComponentSpecificationProperty
Back to top Generated by DocFX