7#include <aws/core/utils/memory/stl/AWSStreamFwd.h>
8#include <aws/core/utils/memory/stl/AWSVector.h>
9#include <aws/ec2/EC2_EXPORTS.h>
10#include <aws/ec2/model/InferenceDeviceInfo.h>
45 template <
typename AcceleratorsT = Aws::Vector<InferenceDeviceInfo>>
47 m_acceleratorsHasBeenSet =
true;
48 m_accelerators = std::forward<AcceleratorsT>(value);
50 template <
typename AcceleratorsT = Aws::Vector<InferenceDeviceInfo>>
55 template <
typename AcceleratorsT = InferenceDeviceInfo>
57 m_acceleratorsHasBeenSet =
true;
58 m_accelerators.emplace_back(std::forward<AcceleratorsT>(value));
71 m_totalInferenceMemoryInMiBHasBeenSet =
true;
72 m_totalInferenceMemoryInMiB = value;
82 int m_totalInferenceMemoryInMiB{0};
83 bool m_acceleratorsHasBeenSet =
false;
84 bool m_totalInferenceMemoryInMiBHasBeenSet =
false;
int GetTotalInferenceMemoryInMiB() const
AWS_EC2_API InferenceAcceleratorInfo(const Aws::Utils::Xml::XmlNode &xmlNode)
bool AcceleratorsHasBeenSet() const
void SetTotalInferenceMemoryInMiB(int value)
InferenceAcceleratorInfo & WithTotalInferenceMemoryInMiB(int value)
void SetAccelerators(AcceleratorsT &&value)
AWS_EC2_API InferenceAcceleratorInfo()=default
AWS_EC2_API void OutputToStream(Aws::OStream &oStream, const char *location) const
InferenceAcceleratorInfo & AddAccelerators(AcceleratorsT &&value)
bool TotalInferenceMemoryInMiBHasBeenSet() const
AWS_EC2_API void OutputToStream(Aws::OStream &ostream, const char *location, unsigned index, const char *locationValue) const
const Aws::Vector< InferenceDeviceInfo > & GetAccelerators() const
AWS_EC2_API InferenceAcceleratorInfo & operator=(const Aws::Utils::Xml::XmlNode &xmlNode)
InferenceAcceleratorInfo & WithAccelerators(AcceleratorsT &&value)
std::vector< T, Aws::Allocator< T > > Vector
std::basic_ostream< char, std::char_traits< char > > OStream