7#include <aws/qconnect/QConnect_EXPORTS.h>
40 m_temperatureHasBeenSet =
true;
41 m_temperature = value;
53 inline double GetTopP()
const {
return m_topP; }
56 m_topPHasBeenSet =
true;
69 inline int GetTopK()
const {
return m_topK; }
72 m_topKHasBeenSet =
true;
88 m_maxTokensToSampleHasBeenSet =
true;
89 m_maxTokensToSample = value;
97 double m_temperature{0.0};
103 int m_maxTokensToSample{0};
104 bool m_temperatureHasBeenSet =
false;
105 bool m_topPHasBeenSet =
false;
106 bool m_topKHasBeenSet =
false;
107 bool m_maxTokensToSampleHasBeenSet =
false;
AWS_QCONNECT_API AIPromptInferenceConfiguration & operator=(Aws::Utils::Json::JsonView jsonValue)
double GetTemperature() const
void SetTopP(double value)
AIPromptInferenceConfiguration & WithTopP(double value)
void SetTemperature(double value)
AWS_QCONNECT_API AIPromptInferenceConfiguration(Aws::Utils::Json::JsonView jsonValue)
AIPromptInferenceConfiguration & WithTopK(int value)
AWS_QCONNECT_API Aws::Utils::Json::JsonValue Jsonize() const
bool TopKHasBeenSet() const
int GetMaxTokensToSample() const
AIPromptInferenceConfiguration & WithMaxTokensToSample(int value)
bool TopPHasBeenSet() const
bool MaxTokensToSampleHasBeenSet() const
void SetMaxTokensToSample(int value)
AWS_QCONNECT_API AIPromptInferenceConfiguration()=default
AIPromptInferenceConfiguration & WithTemperature(double value)
bool TemperatureHasBeenSet() const
Aws::Utils::Json::JsonValue JsonValue