Interface CfnInferenceComponent.InferenceComponentComputeResourceRequirementsProperty
- All Superinterfaces:
software.amazon.jsii.JsiiSerializable
- All Known Implementing Classes:
CfnInferenceComponent.InferenceComponentComputeResourceRequirementsProperty.Jsii$Proxy
- Enclosing class:
CfnInferenceComponent
@Stability(Stable)
public static interface CfnInferenceComponent.InferenceComponentComputeResourceRequirementsProperty
extends software.amazon.jsii.JsiiSerializable
Defines the compute resources to allocate to run a model, plus any adapter models, that you assign to an inference component.
These resources include CPU cores, accelerators, and memory.
Example:
// The code below shows an example of how to instantiate this type.
// The values are placeholders you should change.
import software.amazon.awscdk.services.sagemaker.*;
InferenceComponentComputeResourceRequirementsProperty inferenceComponentComputeResourceRequirementsProperty = InferenceComponentComputeResourceRequirementsProperty.builder()
.maxMemoryRequiredInMb(123)
.minMemoryRequiredInMb(123)
.numberOfAcceleratorDevicesRequired(123)
.numberOfCpuCoresRequired(123)
.build();
- See Also:
-
Nested Class Summary
Nested ClassesModifier and TypeInterfaceDescriptionstatic final classstatic final classAn implementation forCfnInferenceComponent.InferenceComponentComputeResourceRequirementsProperty -
Method Summary
Modifier and TypeMethodDescriptionbuilder()default NumberThe maximum MB of memory to allocate to run a model that you assign to an inference component.default NumberThe minimum MB of memory to allocate to run a model that you assign to an inference component.default NumberThe number of accelerators to allocate to run a model that you assign to an inference component.default NumberThe number of CPU cores to allocate to run a model that you assign to an inference component.Methods inherited from interface software.amazon.jsii.JsiiSerializable
$jsii$toJson
-
Method Details
-
getMaxMemoryRequiredInMb
The maximum MB of memory to allocate to run a model that you assign to an inference component.- See Also:
-
getMinMemoryRequiredInMb
The minimum MB of memory to allocate to run a model that you assign to an inference component.- See Also:
-
getNumberOfAcceleratorDevicesRequired
The number of accelerators to allocate to run a model that you assign to an inference component.Accelerators include GPUs and AWS Inferentia.
- See Also:
-
getNumberOfCpuCoresRequired
The number of CPU cores to allocate to run a model that you assign to an inference component.- See Also:
-
builder
@Stability(Stable) static CfnInferenceComponent.InferenceComponentComputeResourceRequirementsProperty.Builder builder()
-