/AWS1/CL_GLUKAFKASTRMINGSRCO00¶
Additional options for streaming.
CONSTRUCTOR¶
IMPORTING¶
Optional arguments:¶
iv_bootstrapservers TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
A list of bootstrap server URLs, for example, as
b-1.vpc-test-2.o4q88o.c6.kafka.us-east-1.amazonaws.com:9094. This option must be specified in the API call or defined in the table metadata in the Data Catalog.
iv_securityprotocol TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The protocol used to communicate with brokers. The possible values are
"SSL"or"PLAINTEXT".
iv_connectionname TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The name of the connection.
iv_topicname TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The topic name as specified in Apache Kafka. You must specify at least one of
"topicName","assign"or"subscribePattern".
iv_assign TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The specific
TopicPartitionsto consume. You must specify at least one of"topicName","assign"or"subscribePattern".
iv_subscribepattern TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
A Java regex string that identifies the topic list to subscribe to. You must specify at least one of
"topicName","assign"or"subscribePattern".
iv_classification TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
An optional classification.
iv_delimiter TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
Specifies the delimiter character.
iv_startingoffsets TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The starting position in the Kafka topic to read data from. The possible values are
"earliest"or"latest". The default value is"latest".
iv_endingoffsets TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
The end point when a batch query is ended. Possible values are either
"latest"or a JSON string that specifies an ending offset for eachTopicPartition.
iv_polltimeoutms TYPE /AWS1/GLUBOXEDNONNEGATIVELONG /AWS1/GLUBOXEDNONNEGATIVELONG¶
The timeout in milliseconds to poll data from Kafka in Spark job executors. The default value is
512.
iv_numretries TYPE /AWS1/GLUBOXEDNONNEGATIVEINT /AWS1/GLUBOXEDNONNEGATIVEINT¶
The number of times to retry before failing to fetch Kafka offsets. The default value is
3.
iv_retryintervalms TYPE /AWS1/GLUBOXEDNONNEGATIVELONG /AWS1/GLUBOXEDNONNEGATIVELONG¶
The time in milliseconds to wait before retrying to fetch Kafka offsets. The default value is
10.
iv_maxoffsetspertrigger TYPE /AWS1/GLUBOXEDNONNEGATIVELONG /AWS1/GLUBOXEDNONNEGATIVELONG¶
The rate limit on the maximum number of offsets that are processed per trigger interval. The specified total number of offsets is proportionally split across
topicPartitionsof different volumes. The default value is null, which means that the consumer reads all offsets until the known latest offset.
iv_minpartitions TYPE /AWS1/GLUBOXEDNONNEGATIVEINT /AWS1/GLUBOXEDNONNEGATIVEINT¶
The desired minimum number of partitions to read from Kafka. The default value is null, which means that the number of spark partitions is equal to the number of Kafka partitions.
iv_includeheaders TYPE /AWS1/GLUBOXEDBOOLEAN /AWS1/GLUBOXEDBOOLEAN¶
Whether to include the Kafka headers. When the option is set to "true", the data output will contain an additional column named "glue_streaming_kafka_headers" with type
Array[Struct(key: String, value: String)]. The default value is "false". This option is available in Glue version 3.0 or later only.
iv_addrecordtimestamp TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
When this option is set to 'true', the data output will contain an additional column named "__src_timestamp" that indicates the time when the corresponding record received by the topic. The default value is 'false'. This option is supported in Glue version 4.0 or later.
iv_emitconsumerlagmetrics TYPE /AWS1/GLUENCLOSEDINSTRINGPRP /AWS1/GLUENCLOSEDINSTRINGPRP¶
When this option is set to 'true', for each batch, it will emit the metrics for the duration between the oldest record received by the topic and the time it arrives in Glue to CloudWatch. The metric's name is "glue.driver.streaming.maxConsumerLagInMs". The default value is 'false'. This option is supported in Glue version 4.0 or later.
iv_startingtimestamp TYPE /AWS1/GLUISO8601DATETIME /AWS1/GLUISO8601DATETIME¶
The timestamp of the record in the Kafka topic to start reading data from. The possible values are a timestamp string in UTC format of the pattern
yyyy-mm-ddTHH:MM:SSZ(where Z represents a UTC timezone offset with a +/-. For example: "2023-04-04T08:00:00+08:00").Only one of
StartingTimestamporStartingOffsetsmust be set.
Queryable Attributes¶
BootstrapServers¶
A list of bootstrap server URLs, for example, as
b-1.vpc-test-2.o4q88o.c6.kafka.us-east-1.amazonaws.com:9094. This option must be specified in the API call or defined in the table metadata in the Data Catalog.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_BOOTSTRAPSERVERS() |
Getter for BOOTSTRAPSERVERS, with configurable default |
ASK_BOOTSTRAPSERVERS() |
Getter for BOOTSTRAPSERVERS w/ exceptions if field has no va |
HAS_BOOTSTRAPSERVERS() |
Determine if BOOTSTRAPSERVERS has a value |
SecurityProtocol¶
The protocol used to communicate with brokers. The possible values are
"SSL"or"PLAINTEXT".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_SECURITYPROTOCOL() |
Getter for SECURITYPROTOCOL, with configurable default |
ASK_SECURITYPROTOCOL() |
Getter for SECURITYPROTOCOL w/ exceptions if field has no va |
HAS_SECURITYPROTOCOL() |
Determine if SECURITYPROTOCOL has a value |
ConnectionName¶
The name of the connection.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CONNECTIONNAME() |
Getter for CONNECTIONNAME, with configurable default |
ASK_CONNECTIONNAME() |
Getter for CONNECTIONNAME w/ exceptions if field has no valu |
HAS_CONNECTIONNAME() |
Determine if CONNECTIONNAME has a value |
TopicName¶
The topic name as specified in Apache Kafka. You must specify at least one of
"topicName","assign"or"subscribePattern".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_TOPICNAME() |
Getter for TOPICNAME, with configurable default |
ASK_TOPICNAME() |
Getter for TOPICNAME w/ exceptions if field has no value |
HAS_TOPICNAME() |
Determine if TOPICNAME has a value |
Assign¶
The specific
TopicPartitionsto consume. You must specify at least one of"topicName","assign"or"subscribePattern".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ASSIGN() |
Getter for ASSIGN, with configurable default |
ASK_ASSIGN() |
Getter for ASSIGN w/ exceptions if field has no value |
HAS_ASSIGN() |
Determine if ASSIGN has a value |
SubscribePattern¶
A Java regex string that identifies the topic list to subscribe to. You must specify at least one of
"topicName","assign"or"subscribePattern".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_SUBSCRIBEPATTERN() |
Getter for SUBSCRIBEPATTERN, with configurable default |
ASK_SUBSCRIBEPATTERN() |
Getter for SUBSCRIBEPATTERN w/ exceptions if field has no va |
HAS_SUBSCRIBEPATTERN() |
Determine if SUBSCRIBEPATTERN has a value |
Classification¶
An optional classification.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CLASSIFICATION() |
Getter for CLASSIFICATION, with configurable default |
ASK_CLASSIFICATION() |
Getter for CLASSIFICATION w/ exceptions if field has no valu |
HAS_CLASSIFICATION() |
Determine if CLASSIFICATION has a value |
Delimiter¶
Specifies the delimiter character.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DELIMITER() |
Getter for DELIMITER, with configurable default |
ASK_DELIMITER() |
Getter for DELIMITER w/ exceptions if field has no value |
HAS_DELIMITER() |
Determine if DELIMITER has a value |
StartingOffsets¶
The starting position in the Kafka topic to read data from. The possible values are
"earliest"or"latest". The default value is"latest".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_STARTINGOFFSETS() |
Getter for STARTINGOFFSETS, with configurable default |
ASK_STARTINGOFFSETS() |
Getter for STARTINGOFFSETS w/ exceptions if field has no val |
HAS_STARTINGOFFSETS() |
Determine if STARTINGOFFSETS has a value |
EndingOffsets¶
The end point when a batch query is ended. Possible values are either
"latest"or a JSON string that specifies an ending offset for eachTopicPartition.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ENDINGOFFSETS() |
Getter for ENDINGOFFSETS, with configurable default |
ASK_ENDINGOFFSETS() |
Getter for ENDINGOFFSETS w/ exceptions if field has no value |
HAS_ENDINGOFFSETS() |
Determine if ENDINGOFFSETS has a value |
PollTimeoutMs¶
The timeout in milliseconds to poll data from Kafka in Spark job executors. The default value is
512.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_POLLTIMEOUTMS() |
Getter for POLLTIMEOUTMS, with configurable default |
ASK_POLLTIMEOUTMS() |
Getter for POLLTIMEOUTMS w/ exceptions if field has no value |
HAS_POLLTIMEOUTMS() |
Determine if POLLTIMEOUTMS has a value |
NumRetries¶
The number of times to retry before failing to fetch Kafka offsets. The default value is
3.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_NUMRETRIES() |
Getter for NUMRETRIES, with configurable default |
ASK_NUMRETRIES() |
Getter for NUMRETRIES w/ exceptions if field has no value |
HAS_NUMRETRIES() |
Determine if NUMRETRIES has a value |
RetryIntervalMs¶
The time in milliseconds to wait before retrying to fetch Kafka offsets. The default value is
10.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_RETRYINTERVALMS() |
Getter for RETRYINTERVALMS, with configurable default |
ASK_RETRYINTERVALMS() |
Getter for RETRYINTERVALMS w/ exceptions if field has no val |
HAS_RETRYINTERVALMS() |
Determine if RETRYINTERVALMS has a value |
MaxOffsetsPerTrigger¶
The rate limit on the maximum number of offsets that are processed per trigger interval. The specified total number of offsets is proportionally split across
topicPartitionsof different volumes. The default value is null, which means that the consumer reads all offsets until the known latest offset.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_MAXOFFSETSPERTRIGGER() |
Getter for MAXOFFSETSPERTRIGGER, with configurable default |
ASK_MAXOFFSETSPERTRIGGER() |
Getter for MAXOFFSETSPERTRIGGER w/ exceptions if field has n |
HAS_MAXOFFSETSPERTRIGGER() |
Determine if MAXOFFSETSPERTRIGGER has a value |
MinPartitions¶
The desired minimum number of partitions to read from Kafka. The default value is null, which means that the number of spark partitions is equal to the number of Kafka partitions.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_MINPARTITIONS() |
Getter for MINPARTITIONS, with configurable default |
ASK_MINPARTITIONS() |
Getter for MINPARTITIONS w/ exceptions if field has no value |
HAS_MINPARTITIONS() |
Determine if MINPARTITIONS has a value |
IncludeHeaders¶
Whether to include the Kafka headers. When the option is set to "true", the data output will contain an additional column named "glue_streaming_kafka_headers" with type
Array[Struct(key: String, value: String)]. The default value is "false". This option is available in Glue version 3.0 or later only.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_INCLUDEHEADERS() |
Getter for INCLUDEHEADERS, with configurable default |
ASK_INCLUDEHEADERS() |
Getter for INCLUDEHEADERS w/ exceptions if field has no valu |
HAS_INCLUDEHEADERS() |
Determine if INCLUDEHEADERS has a value |
AddRecordTimestamp¶
When this option is set to 'true', the data output will contain an additional column named "__src_timestamp" that indicates the time when the corresponding record received by the topic. The default value is 'false'. This option is supported in Glue version 4.0 or later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ADDRECORDTIMESTAMP() |
Getter for ADDRECORDTIMESTAMP, with configurable default |
ASK_ADDRECORDTIMESTAMP() |
Getter for ADDRECORDTIMESTAMP w/ exceptions if field has no |
HAS_ADDRECORDTIMESTAMP() |
Determine if ADDRECORDTIMESTAMP has a value |
EmitConsumerLagMetrics¶
When this option is set to 'true', for each batch, it will emit the metrics for the duration between the oldest record received by the topic and the time it arrives in Glue to CloudWatch. The metric's name is "glue.driver.streaming.maxConsumerLagInMs". The default value is 'false'. This option is supported in Glue version 4.0 or later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_EMITCONSUMERLAGMETRICS() |
Getter for EMITCONSUMERLAGMETRICS, with configurable default |
ASK_EMITCONSUMERLAGMETRICS() |
Getter for EMITCONSUMERLAGMETRICS w/ exceptions if field has |
HAS_EMITCONSUMERLAGMETRICS() |
Determine if EMITCONSUMERLAGMETRICS has a value |
StartingTimestamp¶
The timestamp of the record in the Kafka topic to start reading data from. The possible values are a timestamp string in UTC format of the pattern
yyyy-mm-ddTHH:MM:SSZ(where Z represents a UTC timezone offset with a +/-. For example: "2023-04-04T08:00:00+08:00").Only one of
StartingTimestamporStartingOffsetsmust be set.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_STARTINGTIMESTAMP() |
Getter for STARTINGTIMESTAMP, with configurable default |
ASK_STARTINGTIMESTAMP() |
Getter for STARTINGTIMESTAMP w/ exceptions if field has no v |
HAS_STARTINGTIMESTAMP() |
Determine if STARTINGTIMESTAMP has a value |