interface InferenceComponentComputeResourceRequirementsProperty
| Language | Type name |
|---|---|
.NET | Amazon.CDK.Mixins.Preview.AWS.SageMaker.Mixins.CfnInferenceComponentPropsMixin.InferenceComponentComputeResourceRequirementsProperty |
Go | github.com/aws/aws-cdk-go/awscdkmixinspreview/v2/awssagemaker/mixins#CfnInferenceComponentPropsMixin_InferenceComponentComputeResourceRequirementsProperty |
Java | software.amazon.awscdk.mixins.preview.services.sagemaker.mixins.CfnInferenceComponentPropsMixin.InferenceComponentComputeResourceRequirementsProperty |
Python | aws_cdk.mixins_preview.aws_sagemaker.mixins.CfnInferenceComponentPropsMixin.InferenceComponentComputeResourceRequirementsProperty |
TypeScript | @aws-cdk/mixins-preview » aws_sagemaker » mixins » CfnInferenceComponentPropsMixin » InferenceComponentComputeResourceRequirementsProperty |
Defines the compute resources to allocate to run a model, plus any adapter models, that you assign to an inference component.
These resources include CPU cores, accelerators, and memory.
Example
// The code below shows an example of how to instantiate this type.
// The values are placeholders you should change.
import { mixins as sagemaker_mixins } from '@aws-cdk/mixins-preview/aws-sagemaker';
const inferenceComponentComputeResourceRequirementsProperty: sagemaker_mixins.CfnInferenceComponentPropsMixin.InferenceComponentComputeResourceRequirementsProperty = {
maxMemoryRequiredInMb: 123,
minMemoryRequiredInMb: 123,
numberOfAcceleratorDevicesRequired: 123,
numberOfCpuCoresRequired: 123,
};
Properties
| Name | Type | Description |
|---|---|---|
| max | number | The maximum MB of memory to allocate to run a model that you assign to an inference component. |
| min | number | The minimum MB of memory to allocate to run a model that you assign to an inference component. |
| number | number | The number of accelerators to allocate to run a model that you assign to an inference component. |
| number | number | The number of CPU cores to allocate to run a model that you assign to an inference component. |
maxMemoryRequiredInMb?
Type:
number
(optional)
The maximum MB of memory to allocate to run a model that you assign to an inference component.
minMemoryRequiredInMb?
Type:
number
(optional)
The minimum MB of memory to allocate to run a model that you assign to an inference component.
numberOfAcceleratorDevicesRequired?
Type:
number
(optional)
The number of accelerators to allocate to run a model that you assign to an inference component.
Accelerators include GPUs and AWS Inferentia.
numberOfCpuCoresRequired?
Type:
number
(optional)
The number of CPU cores to allocate to run a model that you assign to an inference component.

.NET
Go
Java
Python
TypeScript