/AWS1/CL_DMGS3SETTINGS¶
Settings for exporting data to Amazon S3.
CONSTRUCTOR¶
IMPORTING¶
Optional arguments:¶
iv_serviceaccessrolearn TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
The Amazon Resource Name (ARN) used by the service to access the IAM role. The role must allow the
iam:PassRoleaction. It is a required parameter that enables DMS to write and read objects from an S3 bucket.
iv_externaltabledefinition TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
Specifies how tables are defined in the S3 source files only.
iv_csvrowdelimiter TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
The delimiter used to separate rows in the .csv file for both source and target. The default is a carriage return (
\n).
iv_csvdelimiter TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
The delimiter used to separate columns in the .csv file for both source and target. The default is a comma.
iv_bucketfolder TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter isn't specified, then the path used isschema_name/table_name/.
iv_bucketname TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
The name of the S3 bucket.
iv_compressiontype TYPE /AWS1/DMGCOMPRESSIONTYPEVALUE /AWS1/DMGCOMPRESSIONTYPEVALUE¶
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Either set this parameter to NONE (the default) or don't use it to leave the files uncompressed. This parameter applies to both .csv and .parquet file formats.
iv_encryptionmode TYPE /AWS1/DMGENCRYPTIONMODEVALUE /AWS1/DMGENCRYPTIONMODEVALUE¶
The type of server-side encryption that you want to use for your data. This encryption type is part of the endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3(the default) orSSE_KMS.For the
ModifyEndpointoperation, you can change the existing value of theEncryptionModeparameter fromSSE_KMStoSSE_S3. But you can’t change the existing value fromSSE_S3toSSE_KMS.To use
SSE_S3, you need an Identity and Access Management (IAM) role with permission to allow"arn:aws:s3:::dms-*"to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
iv_serversideenckmskeyid TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
If you are using
SSE_KMSfor theEncryptionMode, provide the KMS key ID. The key that you use needs an attached policy that enables Identity and Access Management (IAM) user permissions and allows use of the key.Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
iv_dataformat TYPE /AWS1/DMGDATAFORMATVALUE /AWS1/DMGDATAFORMATVALUE¶
The format of the data that you want to use for output. You can choose one of the following:
csv: This is a row-based file format with comma-separated values (.csv).
parquet: Apache Parquet (.parquet) is a columnar storage file format that features efficient compression and provides faster query response.
iv_encodingtype TYPE /AWS1/DMGENCODINGTYPEVALUE /AWS1/DMGENCODINGTYPEVALUE¶
The type of encoding you are using:
RLE_DICTIONARYuses a combination of bit-packing and run-length encoding to store repeated values more efficiently. This is the default.
PLAINdoesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARYbuilds a dictionary of the values encountered in a given column. The dictionary is stored in a dictionary page for each column chunk.
iv_dictpagesizelimit TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is stored using an encoding type of
PLAIN. This parameter defaults to 1024 * 1024 bytes (1 MiB), the maximum size of a dictionary page before it reverts toPLAINencoding. This size is used for .parquet file format only.
iv_rowgrouplength TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used for .parquet file format only.
If you choose a value larger than the maximum,
RowGroupLengthis set to the max row group length in bytes (64 * 1024 * 1024).
iv_datapagesize TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is used for .parquet file format only.
iv_parquetversion TYPE /AWS1/DMGPARQUETVERSIONVALUE /AWS1/DMGPARQUETVERSIONVALUE¶
The version of the Apache Parquet format that you want to use:
parquet_1_0(the default) orparquet_2_0.
iv_enablestatistics TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
A value that enables statistics for Parquet pages and row groups. Choose
trueto enable statistics,falseto disable. Statistics includeNULL,DISTINCT,MAX, andMINvalues. This parameter defaults totrue. This value is used for .parquet file format only.
iv_includeopforfullload TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) or .parquet output files only to indicate how the rows were added to the source database.
DMS supports the
IncludeOpForFullLoadparameter in versions 3.1.4 and later.DMS supports the use of the .parquet files with the
IncludeOpForFullLoadparameter in versions 3.4.7 and later.For full load, records can only be inserted. By default (the
falsesetting), no information is recorded in these output files for a full load to indicate that the rows were inserted at the source database. IfIncludeOpForFullLoadis set totrueory, the INSERT is recorded as an I annotation in the first field of the .csv file. This allows the format of your target records from a full load to be consistent with the target records from a CDC load.This setting works together with the
CdcInsertsOnlyand theCdcInsertsAndUpdatesparameters for output to .csv files only. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..
iv_cdcinsertsonly TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage (.parquet) output files. By default (the
falsesetting), the first field in a .csv or .parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was inserted, updated, or deleted at the source database for a CDC load to the target.If
CdcInsertsOnlyis set totrueory, only INSERTs from the source database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends on the value ofIncludeOpForFullLoad. IfIncludeOpForFullLoadis set totrue, the first field of every CDC record is set to I to indicate the INSERT operation at the source. IfIncludeOpForFullLoadis set tofalse, every CDC record is written without a first field to indicate the INSERT operation at the source. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..DMS supports the interaction described preceding between the
CdcInsertsOnlyandIncludeOpForFullLoadparameters in versions 3.1.4 and later.
CdcInsertsOnlyandCdcInsertsAndUpdatescan't both be set totruefor the same endpoint. Set eitherCdcInsertsOnlyorCdcInsertsAndUpdatestotruefor the same endpoint, but not both.
iv_timestampcolumnname TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
A value that when nonblank causes DMS to add a column with timestamp information to the endpoint data for an Amazon S3 target.
DMS supports the
TimestampColumnNameparameter in versions 3.1.4 and later.DMS includes an additional
STRINGcolumn in the .csv or .parquet object files of your migrated data when you setTimestampColumnNameto a nonblank value.For a full load, each row of this timestamp column contains a timestamp for when the data was transferred from the source to the target by DMS.
For a change data capture (CDC) load, each row of the timestamp column contains the timestamp for the commit of that row in the source database.
The string format for this timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. By default, the precision of this value is in microseconds. For a CDC load, the rounding of the precision depends on the commit timestamp supported by DMS for the source database.When the
AddColumnNameparameter is set totrue, DMS also includes a name for the timestamp column that you set withTimestampColumnName.
iv_parquettsmpinmillisecond TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
A value that specifies the precision of any
TIMESTAMPcolumn values that are written to an Amazon S3 object file in .parquet format.DMS supports the
ParquetTimestampInMillisecondparameter in versions 3.1.4 and later.When
ParquetTimestampInMillisecondis set totrueory, DMS writes allTIMESTAMPcolumns in a .parquet formatted file with millisecond precision. Otherwise, DMS writes them with microsecond precision.Currently, Amazon Athena and Glue can handle only millisecond precision for
TIMESTAMPvalues. Set this parameter totruefor S3 endpoint object files that are .parquet formatted only if you plan to query or process the data with Athena or Glue.DMS writes any
TIMESTAMPcolumn values written to an S3 file in .csv format with microsecond precision.Setting
ParquetTimestampInMillisecondhas no effect on the string format of the timestamp column value that is inserted by setting theTimestampColumnNameparameter.
iv_cdcinsertsandupdates TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
A value that enables a change data capture (CDC) load to write INSERT and UPDATE operations to .csv or .parquet (columnar storage) output files. The default setting is
false, but whenCdcInsertsAndUpdatesis set totrueory, only INSERTs and UPDATEs from the source database are migrated to the .csv or .parquet file.DMS supports the use of the .parquet files in versions 3.4.7 and later.
How these INSERTs and UPDATEs are recorded depends on the value of the
IncludeOpForFullLoadparameter. IfIncludeOpForFullLoadis set totrue, the first field of every CDC record is set to eitherIorUto indicate INSERT and UPDATE operations at the source. But ifIncludeOpForFullLoadis set tofalse, CDC records are written without an indication of INSERT or UPDATE operations at the source. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..DMS supports the use of the
CdcInsertsAndUpdatesparameter in versions 3.3.1 and later.
CdcInsertsOnlyandCdcInsertsAndUpdatescan't both be set totruefor the same endpoint. Set eitherCdcInsertsOnlyorCdcInsertsAndUpdatestotruefor the same endpoint, but not both.
iv_datepartitionenabled TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
When set to
true, this parameter partitions S3 bucket folders based on transaction commit dates. The default value isfalse. For more information about date-based folder partitioning, see Using date-based folder partitioning.
iv_datepartitionsequence TYPE /AWS1/DMGDATEPARTITIONSEQUEN00 /AWS1/DMGDATEPARTITIONSEQUEN00¶
Identifies the sequence of the date format to use during folder partitioning. The default value is
YYYYMMDD. Use this parameter whenDatePartitionedEnabledis set totrue.
iv_datepartitiondelimiter TYPE /AWS1/DMGDATEPARTITIONDELMTE00 /AWS1/DMGDATEPARTITIONDELMTE00¶
Specifies a date separating delimiter to use during folder partitioning. The default value is
SLASH. Use this parameter whenDatePartitionedEnabledis set totrue.
iv_usecsvnosupvalue TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
This setting applies if the S3 output files during a change data capture (CDC) load are written in .csv format. If set to
truefor columns not included in the supplemental log, DMS uses the value specified byCsvNoSupValue. If not set or set tofalse, DMS uses the null value for these columns.This setting is supported in DMS versions 3.4.1 and later.
iv_csvnosupvalue TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
This setting only applies if your Amazon S3 output files during a change data capture (CDC) load are written in .csv format. If
UseCsvNoSupValueis set to true, specify a string value that you want DMS to use for all columns not included in the supplemental log. If you do not specify a string value, DMS uses the null value for these columns regardless of theUseCsvNoSupValuesetting.This setting is supported in DMS versions 3.4.1 and later.
iv_preservetransactions TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
If set to
true, DMS saves the transaction order for a change data capture (CDC) load on the Amazon S3 target specified byCdcPath. For more information, see Capturing data changes (CDC) including transaction order on the S3 target.This setting is supported in DMS versions 3.4.2 and later.
iv_cdcpath TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
Specifies the folder path of CDC files. For an S3 source, this setting is required if a task captures change data; otherwise, it's optional. If
CdcPathis set, DMS reads CDC files from this path and replicates the data changes to the target endpoint. For an S3 target if you setPreserveTransactionstotrue, DMS verifies that you have set this parameter to a folder path on your S3 target where DMS can save the transaction order for the CDC load. DMS creates this CDC folder path in either your S3 target working directory or the S3 target location specified byBucketFolderandBucketName.For example, if you specify
CdcPathasMyChangedData, and you specifyBucketNameasMyTargetBucketbut do not specifyBucketFolder, DMS creates the CDC folder path following:MyTargetBucket/MyChangedData.If you specify the same
CdcPath, and you specifyBucketNameasMyTargetBucketandBucketFolderasMyTargetData, DMS creates the CDC folder path following:MyTargetBucket/MyTargetData/MyChangedData.For more information on CDC including transaction order on an S3 target, see Capturing data changes (CDC) including transaction order on the S3 target.
This setting is supported in DMS versions 3.4.2 and later.
iv_usetskstrttimeforfulllo00 TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
When set to true, this parameter uses the task start time as the timestamp column value instead of the time data is written to target. For full load, when
useTaskStartTimeForFullLoadTimestampis set totrue, each row of the timestamp column contains the task start time. For CDC loads, each row of the timestamp column contains the transaction commit time.When
useTaskStartTimeForFullLoadTimestampis set tofalse, the full load timestamp in the timestamp column increments with the time data arrives at the target.
iv_cannedaclforobjects TYPE /AWS1/DMGCANNEDACLFOROBJSVALUE /AWS1/DMGCANNEDACLFOROBJSVALUE¶
A value that enables DMS to specify a predefined (canned) access control list for objects created in an Amazon S3 bucket as .csv or .parquet files. For more information about Amazon S3 canned ACLs, see Canned ACL in the Amazon S3 Developer Guide.
The default value is NONE. Valid values include NONE, PRIVATE, PUBLIC_READ, PUBLIC_READ_WRITE, AUTHENTICATED_READ, AWS_EXEC_READ, BUCKET_OWNER_READ, and BUCKET_OWNER_FULL_CONTROL.
iv_addcolumnname TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
An optional parameter that, when set to
trueory, you can use to add column name information to the .csv output file.The default value is
false. Valid values aretrue,false,y, andn.
iv_cdcmaxbatchinterval TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
Maximum length of the interval, defined in seconds, after which to output a file to Amazon S3.
When
CdcMaxBatchIntervalandCdcMinFileSizeare both specified, the file write is triggered by whichever parameter condition is met first within an DMS CloudFormation template.The default value is 60 seconds.
iv_cdcminfilesize TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
Minimum file size, defined in kilobytes, to reach for a file output to Amazon S3.
When
CdcMinFileSizeandCdcMaxBatchIntervalare both specified, the file write is triggered by whichever parameter condition is met first within an DMS CloudFormation template.The default value is 32 MB.
iv_csvnullvalue TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
An optional parameter that specifies how DMS treats null values. While handling the null value, you can use this parameter to pass a user-defined string as null when writing to the target. For example, when target columns are nullable, you can use this option to differentiate between the empty string value and the null value. So, if you set this parameter value to the empty string ("" or ''), DMS treats the empty string as the null value instead of
NULL.The default value is
NULL. Valid values include any valid string.
iv_ignoreheaderrows TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
When this value is set to 1, DMS ignores the first row header in a .csv file. A value of 1 turns on the feature; a value of 0 turns off the feature.
The default is 0.
iv_maxfilesize TYPE /AWS1/DMGINTEGEROPTIONAL /AWS1/DMGINTEGEROPTIONAL¶
A value that specifies the maximum size (in KB) of any .csv file to be created while migrating to an S3 target during full load.
The default value is 1,048,576 KB (1 GB). Valid values include 1 to 1,048,576.
iv_rfc4180 TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
For an S3 source, when this value is set to
trueory, each leading double quotation mark has to be followed by an ending double quotation mark. This formatting complies with RFC 4180. When this value is set tofalseorn, string literals are copied to the target as is. In this case, a delimiter (row or column) signals the end of the field. Thus, you can't use a delimiter as part of the string, because it signals the end of the value.For an S3 target, an optional parameter used to set behavior to comply with RFC 4180 for data migrated to Amazon S3 using .csv file format only. When this value is set to
trueoryusing Amazon S3 as a target, if the data has quotation marks or newline characters in it, DMS encloses the entire column with an additional pair of double quotation marks ("). Every quotation mark within the data is repeated twice.The default value is
true. Valid values includetrue,false,y, andn.
iv_datepartitiontimezone TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
When creating an S3 target endpoint, set
DatePartitionTimezoneto convert the current UTC time into a specified time zone. The conversion occurs when a date partition folder is created and a CDC filename is generated. The time zone format is Area/Location. Use this parameter whenDatePartitionedEnabledis set to true, as shown in the following example:
s3-settings='{"DatePartitionEnabled": true, "DatePartitionSequence": "YYYYMMDDHH", "DatePartitionDelimiter": "SLASH", "DatePartitionTimezone":"Asia/Seoul", "BucketName": "dms-nattarat-test"}'
iv_addtrailingpaddingchara00 TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
Use the S3 target endpoint setting
AddTrailingPaddingCharacterto add padding on string data. The default value isfalse.
iv_expectedbucketowner TYPE /AWS1/DMGSTRING /AWS1/DMGSTRING¶
To specify a bucket owner and prevent sniping, you can use the
ExpectedBucketOwnerendpoint setting.Example:
--s3-settings='{"ExpectedBucketOwner": "AWS_Account_ID"}'When you make a request to test a connection or perform a migration, S3 checks the account ID of the bucket owner against the specified parameter.
iv_gluecataloggeneration TYPE /AWS1/DMGBOOLEANOPTIONAL /AWS1/DMGBOOLEANOPTIONAL¶
When true, allows Glue to catalog your S3 bucket. Creating an Glue catalog lets you use Athena to query your data.
Queryable Attributes¶
ServiceAccessRoleArn¶
The Amazon Resource Name (ARN) used by the service to access the IAM role. The role must allow the
iam:PassRoleaction. It is a required parameter that enables DMS to write and read objects from an S3 bucket.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_SERVICEACCESSROLEARN() |
Getter for SERVICEACCESSROLEARN, with configurable default |
ASK_SERVICEACCESSROLEARN() |
Getter for SERVICEACCESSROLEARN w/ exceptions if field has n |
HAS_SERVICEACCESSROLEARN() |
Determine if SERVICEACCESSROLEARN has a value |
ExternalTableDefinition¶
Specifies how tables are defined in the S3 source files only.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_EXTERNALTABLEDEFINITION() |
Getter for EXTERNALTABLEDEFINITION, with configurable defaul |
ASK_EXTERNALTABLEDEFINITION() |
Getter for EXTERNALTABLEDEFINITION w/ exceptions if field ha |
HAS_EXTERNALTABLEDEFINITION() |
Determine if EXTERNALTABLEDEFINITION has a value |
CsvRowDelimiter¶
The delimiter used to separate rows in the .csv file for both source and target. The default is a carriage return (
\n).
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CSVROWDELIMITER() |
Getter for CSVROWDELIMITER, with configurable default |
ASK_CSVROWDELIMITER() |
Getter for CSVROWDELIMITER w/ exceptions if field has no val |
HAS_CSVROWDELIMITER() |
Determine if CSVROWDELIMITER has a value |
CsvDelimiter¶
The delimiter used to separate columns in the .csv file for both source and target. The default is a comma.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CSVDELIMITER() |
Getter for CSVDELIMITER, with configurable default |
ASK_CSVDELIMITER() |
Getter for CSVDELIMITER w/ exceptions if field has no value |
HAS_CSVDELIMITER() |
Determine if CSVDELIMITER has a value |
BucketFolder¶
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter isn't specified, then the path used isschema_name/table_name/.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_BUCKETFOLDER() |
Getter for BUCKETFOLDER, with configurable default |
ASK_BUCKETFOLDER() |
Getter for BUCKETFOLDER w/ exceptions if field has no value |
HAS_BUCKETFOLDER() |
Determine if BUCKETFOLDER has a value |
BucketName¶
The name of the S3 bucket.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_BUCKETNAME() |
Getter for BUCKETNAME, with configurable default |
ASK_BUCKETNAME() |
Getter for BUCKETNAME w/ exceptions if field has no value |
HAS_BUCKETNAME() |
Determine if BUCKETNAME has a value |
CompressionType¶
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Either set this parameter to NONE (the default) or don't use it to leave the files uncompressed. This parameter applies to both .csv and .parquet file formats.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_COMPRESSIONTYPE() |
Getter for COMPRESSIONTYPE, with configurable default |
ASK_COMPRESSIONTYPE() |
Getter for COMPRESSIONTYPE w/ exceptions if field has no val |
HAS_COMPRESSIONTYPE() |
Determine if COMPRESSIONTYPE has a value |
EncryptionMode¶
The type of server-side encryption that you want to use for your data. This encryption type is part of the endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3(the default) orSSE_KMS.For the
ModifyEndpointoperation, you can change the existing value of theEncryptionModeparameter fromSSE_KMStoSSE_S3. But you can’t change the existing value fromSSE_S3toSSE_KMS.To use
SSE_S3, you need an Identity and Access Management (IAM) role with permission to allow"arn:aws:s3:::dms-*"to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ENCRYPTIONMODE() |
Getter for ENCRYPTIONMODE, with configurable default |
ASK_ENCRYPTIONMODE() |
Getter for ENCRYPTIONMODE w/ exceptions if field has no valu |
HAS_ENCRYPTIONMODE() |
Determine if ENCRYPTIONMODE has a value |
ServerSideEncryptionKmsKeyId¶
If you are using
SSE_KMSfor theEncryptionMode, provide the KMS key ID. The key that you use needs an attached policy that enables Identity and Access Management (IAM) user permissions and allows use of the key.Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_SERVERSIDEENCKMSKEYID() |
Getter for SERVERSIDEENCRYPTIONKMSKEYID, with configurable d |
ASK_SERVERSIDEENCKMSKEYID() |
Getter for SERVERSIDEENCRYPTIONKMSKEYID w/ exceptions if fie |
HAS_SERVERSIDEENCKMSKEYID() |
Determine if SERVERSIDEENCRYPTIONKMSKEYID has a value |
DataFormat¶
The format of the data that you want to use for output. You can choose one of the following:
csv: This is a row-based file format with comma-separated values (.csv).
parquet: Apache Parquet (.parquet) is a columnar storage file format that features efficient compression and provides faster query response.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATAFORMAT() |
Getter for DATAFORMAT, with configurable default |
ASK_DATAFORMAT() |
Getter for DATAFORMAT w/ exceptions if field has no value |
HAS_DATAFORMAT() |
Determine if DATAFORMAT has a value |
EncodingType¶
The type of encoding you are using:
RLE_DICTIONARYuses a combination of bit-packing and run-length encoding to store repeated values more efficiently. This is the default.
PLAINdoesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARYbuilds a dictionary of the values encountered in a given column. The dictionary is stored in a dictionary page for each column chunk.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ENCODINGTYPE() |
Getter for ENCODINGTYPE, with configurable default |
ASK_ENCODINGTYPE() |
Getter for ENCODINGTYPE w/ exceptions if field has no value |
HAS_ENCODINGTYPE() |
Determine if ENCODINGTYPE has a value |
DictPageSizeLimit¶
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is stored using an encoding type of
PLAIN. This parameter defaults to 1024 * 1024 bytes (1 MiB), the maximum size of a dictionary page before it reverts toPLAINencoding. This size is used for .parquet file format only.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DICTPAGESIZELIMIT() |
Getter for DICTPAGESIZELIMIT, with configurable default |
ASK_DICTPAGESIZELIMIT() |
Getter for DICTPAGESIZELIMIT w/ exceptions if field has no v |
HAS_DICTPAGESIZELIMIT() |
Determine if DICTPAGESIZELIMIT has a value |
RowGroupLength¶
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used for .parquet file format only.
If you choose a value larger than the maximum,
RowGroupLengthis set to the max row group length in bytes (64 * 1024 * 1024).
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ROWGROUPLENGTH() |
Getter for ROWGROUPLENGTH, with configurable default |
ASK_ROWGROUPLENGTH() |
Getter for ROWGROUPLENGTH w/ exceptions if field has no valu |
HAS_ROWGROUPLENGTH() |
Determine if ROWGROUPLENGTH has a value |
DataPageSize¶
The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is used for .parquet file format only.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATAPAGESIZE() |
Getter for DATAPAGESIZE, with configurable default |
ASK_DATAPAGESIZE() |
Getter for DATAPAGESIZE w/ exceptions if field has no value |
HAS_DATAPAGESIZE() |
Determine if DATAPAGESIZE has a value |
ParquetVersion¶
The version of the Apache Parquet format that you want to use:
parquet_1_0(the default) orparquet_2_0.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_PARQUETVERSION() |
Getter for PARQUETVERSION, with configurable default |
ASK_PARQUETVERSION() |
Getter for PARQUETVERSION w/ exceptions if field has no valu |
HAS_PARQUETVERSION() |
Determine if PARQUETVERSION has a value |
EnableStatistics¶
A value that enables statistics for Parquet pages and row groups. Choose
trueto enable statistics,falseto disable. Statistics includeNULL,DISTINCT,MAX, andMINvalues. This parameter defaults totrue. This value is used for .parquet file format only.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ENABLESTATISTICS() |
Getter for ENABLESTATISTICS, with configurable default |
ASK_ENABLESTATISTICS() |
Getter for ENABLESTATISTICS w/ exceptions if field has no va |
HAS_ENABLESTATISTICS() |
Determine if ENABLESTATISTICS has a value |
IncludeOpForFullLoad¶
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) or .parquet output files only to indicate how the rows were added to the source database.
DMS supports the
IncludeOpForFullLoadparameter in versions 3.1.4 and later.DMS supports the use of the .parquet files with the
IncludeOpForFullLoadparameter in versions 3.4.7 and later.For full load, records can only be inserted. By default (the
falsesetting), no information is recorded in these output files for a full load to indicate that the rows were inserted at the source database. IfIncludeOpForFullLoadis set totrueory, the INSERT is recorded as an I annotation in the first field of the .csv file. This allows the format of your target records from a full load to be consistent with the target records from a CDC load.This setting works together with the
CdcInsertsOnlyand theCdcInsertsAndUpdatesparameters for output to .csv files only. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_INCLUDEOPFORFULLLOAD() |
Getter for INCLUDEOPFORFULLLOAD, with configurable default |
ASK_INCLUDEOPFORFULLLOAD() |
Getter for INCLUDEOPFORFULLLOAD w/ exceptions if field has n |
HAS_INCLUDEOPFORFULLLOAD() |
Determine if INCLUDEOPFORFULLLOAD has a value |
CdcInsertsOnly¶
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage (.parquet) output files. By default (the
falsesetting), the first field in a .csv or .parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was inserted, updated, or deleted at the source database for a CDC load to the target.If
CdcInsertsOnlyis set totrueory, only INSERTs from the source database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends on the value ofIncludeOpForFullLoad. IfIncludeOpForFullLoadis set totrue, the first field of every CDC record is set to I to indicate the INSERT operation at the source. IfIncludeOpForFullLoadis set tofalse, every CDC record is written without a first field to indicate the INSERT operation at the source. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..DMS supports the interaction described preceding between the
CdcInsertsOnlyandIncludeOpForFullLoadparameters in versions 3.1.4 and later.
CdcInsertsOnlyandCdcInsertsAndUpdatescan't both be set totruefor the same endpoint. Set eitherCdcInsertsOnlyorCdcInsertsAndUpdatestotruefor the same endpoint, but not both.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CDCINSERTSONLY() |
Getter for CDCINSERTSONLY, with configurable default |
ASK_CDCINSERTSONLY() |
Getter for CDCINSERTSONLY w/ exceptions if field has no valu |
HAS_CDCINSERTSONLY() |
Determine if CDCINSERTSONLY has a value |
TimestampColumnName¶
A value that when nonblank causes DMS to add a column with timestamp information to the endpoint data for an Amazon S3 target.
DMS supports the
TimestampColumnNameparameter in versions 3.1.4 and later.DMS includes an additional
STRINGcolumn in the .csv or .parquet object files of your migrated data when you setTimestampColumnNameto a nonblank value.For a full load, each row of this timestamp column contains a timestamp for when the data was transferred from the source to the target by DMS.
For a change data capture (CDC) load, each row of the timestamp column contains the timestamp for the commit of that row in the source database.
The string format for this timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. By default, the precision of this value is in microseconds. For a CDC load, the rounding of the precision depends on the commit timestamp supported by DMS for the source database.When the
AddColumnNameparameter is set totrue, DMS also includes a name for the timestamp column that you set withTimestampColumnName.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_TIMESTAMPCOLUMNNAME() |
Getter for TIMESTAMPCOLUMNNAME, with configurable default |
ASK_TIMESTAMPCOLUMNNAME() |
Getter for TIMESTAMPCOLUMNNAME w/ exceptions if field has no |
HAS_TIMESTAMPCOLUMNNAME() |
Determine if TIMESTAMPCOLUMNNAME has a value |
ParquetTimestampInMillisecond¶
A value that specifies the precision of any
TIMESTAMPcolumn values that are written to an Amazon S3 object file in .parquet format.DMS supports the
ParquetTimestampInMillisecondparameter in versions 3.1.4 and later.When
ParquetTimestampInMillisecondis set totrueory, DMS writes allTIMESTAMPcolumns in a .parquet formatted file with millisecond precision. Otherwise, DMS writes them with microsecond precision.Currently, Amazon Athena and Glue can handle only millisecond precision for
TIMESTAMPvalues. Set this parameter totruefor S3 endpoint object files that are .parquet formatted only if you plan to query or process the data with Athena or Glue.DMS writes any
TIMESTAMPcolumn values written to an S3 file in .csv format with microsecond precision.Setting
ParquetTimestampInMillisecondhas no effect on the string format of the timestamp column value that is inserted by setting theTimestampColumnNameparameter.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_PARQUETTSMPINMILLISECOND() |
Getter for PARQUETTSMPINMILLISECOND, with configurable defau |
ASK_PARQUETTSMPINMILLISECOND() |
Getter for PARQUETTSMPINMILLISECOND w/ exceptions if field h |
HAS_PARQUETTSMPINMILLISECOND() |
Determine if PARQUETTSMPINMILLISECOND has a value |
CdcInsertsAndUpdates¶
A value that enables a change data capture (CDC) load to write INSERT and UPDATE operations to .csv or .parquet (columnar storage) output files. The default setting is
false, but whenCdcInsertsAndUpdatesis set totrueory, only INSERTs and UPDATEs from the source database are migrated to the .csv or .parquet file.DMS supports the use of the .parquet files in versions 3.4.7 and later.
How these INSERTs and UPDATEs are recorded depends on the value of the
IncludeOpForFullLoadparameter. IfIncludeOpForFullLoadis set totrue, the first field of every CDC record is set to eitherIorUto indicate INSERT and UPDATE operations at the source. But ifIncludeOpForFullLoadis set tofalse, CDC records are written without an indication of INSERT or UPDATE operations at the source. For more information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the Database Migration Service User Guide..DMS supports the use of the
CdcInsertsAndUpdatesparameter in versions 3.3.1 and later.
CdcInsertsOnlyandCdcInsertsAndUpdatescan't both be set totruefor the same endpoint. Set eitherCdcInsertsOnlyorCdcInsertsAndUpdatestotruefor the same endpoint, but not both.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CDCINSERTSANDUPDATES() |
Getter for CDCINSERTSANDUPDATES, with configurable default |
ASK_CDCINSERTSANDUPDATES() |
Getter for CDCINSERTSANDUPDATES w/ exceptions if field has n |
HAS_CDCINSERTSANDUPDATES() |
Determine if CDCINSERTSANDUPDATES has a value |
DatePartitionEnabled¶
When set to
true, this parameter partitions S3 bucket folders based on transaction commit dates. The default value isfalse. For more information about date-based folder partitioning, see Using date-based folder partitioning.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATEPARTITIONENABLED() |
Getter for DATEPARTITIONENABLED, with configurable default |
ASK_DATEPARTITIONENABLED() |
Getter for DATEPARTITIONENABLED w/ exceptions if field has n |
HAS_DATEPARTITIONENABLED() |
Determine if DATEPARTITIONENABLED has a value |
DatePartitionSequence¶
Identifies the sequence of the date format to use during folder partitioning. The default value is
YYYYMMDD. Use this parameter whenDatePartitionedEnabledis set totrue.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATEPARTITIONSEQUENCE() |
Getter for DATEPARTITIONSEQUENCE, with configurable default |
ASK_DATEPARTITIONSEQUENCE() |
Getter for DATEPARTITIONSEQUENCE w/ exceptions if field has |
HAS_DATEPARTITIONSEQUENCE() |
Determine if DATEPARTITIONSEQUENCE has a value |
DatePartitionDelimiter¶
Specifies a date separating delimiter to use during folder partitioning. The default value is
SLASH. Use this parameter whenDatePartitionedEnabledis set totrue.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATEPARTITIONDELIMITER() |
Getter for DATEPARTITIONDELIMITER, with configurable default |
ASK_DATEPARTITIONDELIMITER() |
Getter for DATEPARTITIONDELIMITER w/ exceptions if field has |
HAS_DATEPARTITIONDELIMITER() |
Determine if DATEPARTITIONDELIMITER has a value |
UseCsvNoSupValue¶
This setting applies if the S3 output files during a change data capture (CDC) load are written in .csv format. If set to
truefor columns not included in the supplemental log, DMS uses the value specified byCsvNoSupValue. If not set or set tofalse, DMS uses the null value for these columns.This setting is supported in DMS versions 3.4.1 and later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_USECSVNOSUPVALUE() |
Getter for USECSVNOSUPVALUE, with configurable default |
ASK_USECSVNOSUPVALUE() |
Getter for USECSVNOSUPVALUE w/ exceptions if field has no va |
HAS_USECSVNOSUPVALUE() |
Determine if USECSVNOSUPVALUE has a value |
CsvNoSupValue¶
This setting only applies if your Amazon S3 output files during a change data capture (CDC) load are written in .csv format. If
UseCsvNoSupValueis set to true, specify a string value that you want DMS to use for all columns not included in the supplemental log. If you do not specify a string value, DMS uses the null value for these columns regardless of theUseCsvNoSupValuesetting.This setting is supported in DMS versions 3.4.1 and later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CSVNOSUPVALUE() |
Getter for CSVNOSUPVALUE, with configurable default |
ASK_CSVNOSUPVALUE() |
Getter for CSVNOSUPVALUE w/ exceptions if field has no value |
HAS_CSVNOSUPVALUE() |
Determine if CSVNOSUPVALUE has a value |
PreserveTransactions¶
If set to
true, DMS saves the transaction order for a change data capture (CDC) load on the Amazon S3 target specified byCdcPath. For more information, see Capturing data changes (CDC) including transaction order on the S3 target.This setting is supported in DMS versions 3.4.2 and later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_PRESERVETRANSACTIONS() |
Getter for PRESERVETRANSACTIONS, with configurable default |
ASK_PRESERVETRANSACTIONS() |
Getter for PRESERVETRANSACTIONS w/ exceptions if field has n |
HAS_PRESERVETRANSACTIONS() |
Determine if PRESERVETRANSACTIONS has a value |
CdcPath¶
Specifies the folder path of CDC files. For an S3 source, this setting is required if a task captures change data; otherwise, it's optional. If
CdcPathis set, DMS reads CDC files from this path and replicates the data changes to the target endpoint. For an S3 target if you setPreserveTransactionstotrue, DMS verifies that you have set this parameter to a folder path on your S3 target where DMS can save the transaction order for the CDC load. DMS creates this CDC folder path in either your S3 target working directory or the S3 target location specified byBucketFolderandBucketName.For example, if you specify
CdcPathasMyChangedData, and you specifyBucketNameasMyTargetBucketbut do not specifyBucketFolder, DMS creates the CDC folder path following:MyTargetBucket/MyChangedData.If you specify the same
CdcPath, and you specifyBucketNameasMyTargetBucketandBucketFolderasMyTargetData, DMS creates the CDC folder path following:MyTargetBucket/MyTargetData/MyChangedData.For more information on CDC including transaction order on an S3 target, see Capturing data changes (CDC) including transaction order on the S3 target.
This setting is supported in DMS versions 3.4.2 and later.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CDCPATH() |
Getter for CDCPATH, with configurable default |
ASK_CDCPATH() |
Getter for CDCPATH w/ exceptions if field has no value |
HAS_CDCPATH() |
Determine if CDCPATH has a value |
UseTaskStartTimeForFullLoadTimestamp¶
When set to true, this parameter uses the task start time as the timestamp column value instead of the time data is written to target. For full load, when
useTaskStartTimeForFullLoadTimestampis set totrue, each row of the timestamp column contains the task start time. For CDC loads, each row of the timestamp column contains the transaction commit time.When
useTaskStartTimeForFullLoadTimestampis set tofalse, the full load timestamp in the timestamp column increments with the time data arrives at the target.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_USETSKSTRTTIMEFORFULLL00() |
Getter for USETASKSTRTTIMEFORFULLLOADTS, with configurable d |
ASK_USETSKSTRTTIMEFORFULLL00() |
Getter for USETASKSTRTTIMEFORFULLLOADTS w/ exceptions if fie |
HAS_USETSKSTRTTIMEFORFULLL00() |
Determine if USETASKSTRTTIMEFORFULLLOADTS has a value |
CannedAclForObjects¶
A value that enables DMS to specify a predefined (canned) access control list for objects created in an Amazon S3 bucket as .csv or .parquet files. For more information about Amazon S3 canned ACLs, see Canned ACL in the Amazon S3 Developer Guide.
The default value is NONE. Valid values include NONE, PRIVATE, PUBLIC_READ, PUBLIC_READ_WRITE, AUTHENTICATED_READ, AWS_EXEC_READ, BUCKET_OWNER_READ, and BUCKET_OWNER_FULL_CONTROL.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CANNEDACLFOROBJECTS() |
Getter for CANNEDACLFOROBJECTS, with configurable default |
ASK_CANNEDACLFOROBJECTS() |
Getter for CANNEDACLFOROBJECTS w/ exceptions if field has no |
HAS_CANNEDACLFOROBJECTS() |
Determine if CANNEDACLFOROBJECTS has a value |
AddColumnName¶
An optional parameter that, when set to
trueory, you can use to add column name information to the .csv output file.The default value is
false. Valid values aretrue,false,y, andn.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ADDCOLUMNNAME() |
Getter for ADDCOLUMNNAME, with configurable default |
ASK_ADDCOLUMNNAME() |
Getter for ADDCOLUMNNAME w/ exceptions if field has no value |
HAS_ADDCOLUMNNAME() |
Determine if ADDCOLUMNNAME has a value |
CdcMaxBatchInterval¶
Maximum length of the interval, defined in seconds, after which to output a file to Amazon S3.
When
CdcMaxBatchIntervalandCdcMinFileSizeare both specified, the file write is triggered by whichever parameter condition is met first within an DMS CloudFormation template.The default value is 60 seconds.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CDCMAXBATCHINTERVAL() |
Getter for CDCMAXBATCHINTERVAL, with configurable default |
ASK_CDCMAXBATCHINTERVAL() |
Getter for CDCMAXBATCHINTERVAL w/ exceptions if field has no |
HAS_CDCMAXBATCHINTERVAL() |
Determine if CDCMAXBATCHINTERVAL has a value |
CdcMinFileSize¶
Minimum file size, defined in kilobytes, to reach for a file output to Amazon S3.
When
CdcMinFileSizeandCdcMaxBatchIntervalare both specified, the file write is triggered by whichever parameter condition is met first within an DMS CloudFormation template.The default value is 32 MB.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CDCMINFILESIZE() |
Getter for CDCMINFILESIZE, with configurable default |
ASK_CDCMINFILESIZE() |
Getter for CDCMINFILESIZE w/ exceptions if field has no valu |
HAS_CDCMINFILESIZE() |
Determine if CDCMINFILESIZE has a value |
CsvNullValue¶
An optional parameter that specifies how DMS treats null values. While handling the null value, you can use this parameter to pass a user-defined string as null when writing to the target. For example, when target columns are nullable, you can use this option to differentiate between the empty string value and the null value. So, if you set this parameter value to the empty string ("" or ''), DMS treats the empty string as the null value instead of
NULL.The default value is
NULL. Valid values include any valid string.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CSVNULLVALUE() |
Getter for CSVNULLVALUE, with configurable default |
ASK_CSVNULLVALUE() |
Getter for CSVNULLVALUE w/ exceptions if field has no value |
HAS_CSVNULLVALUE() |
Determine if CSVNULLVALUE has a value |
IgnoreHeaderRows¶
When this value is set to 1, DMS ignores the first row header in a .csv file. A value of 1 turns on the feature; a value of 0 turns off the feature.
The default is 0.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_IGNOREHEADERROWS() |
Getter for IGNOREHEADERROWS, with configurable default |
ASK_IGNOREHEADERROWS() |
Getter for IGNOREHEADERROWS w/ exceptions if field has no va |
HAS_IGNOREHEADERROWS() |
Determine if IGNOREHEADERROWS has a value |
MaxFileSize¶
A value that specifies the maximum size (in KB) of any .csv file to be created while migrating to an S3 target during full load.
The default value is 1,048,576 KB (1 GB). Valid values include 1 to 1,048,576.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_MAXFILESIZE() |
Getter for MAXFILESIZE, with configurable default |
ASK_MAXFILESIZE() |
Getter for MAXFILESIZE w/ exceptions if field has no value |
HAS_MAXFILESIZE() |
Determine if MAXFILESIZE has a value |
Rfc4180¶
For an S3 source, when this value is set to
trueory, each leading double quotation mark has to be followed by an ending double quotation mark. This formatting complies with RFC 4180. When this value is set tofalseorn, string literals are copied to the target as is. In this case, a delimiter (row or column) signals the end of the field. Thus, you can't use a delimiter as part of the string, because it signals the end of the value.For an S3 target, an optional parameter used to set behavior to comply with RFC 4180 for data migrated to Amazon S3 using .csv file format only. When this value is set to
trueoryusing Amazon S3 as a target, if the data has quotation marks or newline characters in it, DMS encloses the entire column with an additional pair of double quotation marks ("). Every quotation mark within the data is repeated twice.The default value is
true. Valid values includetrue,false,y, andn.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_RFC4180() |
Getter for RFC4180, with configurable default |
ASK_RFC4180() |
Getter for RFC4180 w/ exceptions if field has no value |
HAS_RFC4180() |
Determine if RFC4180 has a value |
DatePartitionTimezone¶
When creating an S3 target endpoint, set
DatePartitionTimezoneto convert the current UTC time into a specified time zone. The conversion occurs when a date partition folder is created and a CDC filename is generated. The time zone format is Area/Location. Use this parameter whenDatePartitionedEnabledis set to true, as shown in the following example:
s3-settings='{"DatePartitionEnabled": true, "DatePartitionSequence": "YYYYMMDDHH", "DatePartitionDelimiter": "SLASH", "DatePartitionTimezone":"Asia/Seoul", "BucketName": "dms-nattarat-test"}'
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_DATEPARTITIONTIMEZONE() |
Getter for DATEPARTITIONTIMEZONE, with configurable default |
ASK_DATEPARTITIONTIMEZONE() |
Getter for DATEPARTITIONTIMEZONE w/ exceptions if field has |
HAS_DATEPARTITIONTIMEZONE() |
Determine if DATEPARTITIONTIMEZONE has a value |
AddTrailingPaddingCharacter¶
Use the S3 target endpoint setting
AddTrailingPaddingCharacterto add padding on string data. The default value isfalse.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_ADDTRAILINGPADDINGCHAR00() |
Getter for ADDTRAILINGPADDINGCHARACTER, with configurable de |
ASK_ADDTRAILINGPADDINGCHAR00() |
Getter for ADDTRAILINGPADDINGCHARACTER w/ exceptions if fiel |
HAS_ADDTRAILINGPADDINGCHAR00() |
Determine if ADDTRAILINGPADDINGCHARACTER has a value |
ExpectedBucketOwner¶
To specify a bucket owner and prevent sniping, you can use the
ExpectedBucketOwnerendpoint setting.Example:
--s3-settings='{"ExpectedBucketOwner": "AWS_Account_ID"}'When you make a request to test a connection or perform a migration, S3 checks the account ID of the bucket owner against the specified parameter.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_EXPECTEDBUCKETOWNER() |
Getter for EXPECTEDBUCKETOWNER, with configurable default |
ASK_EXPECTEDBUCKETOWNER() |
Getter for EXPECTEDBUCKETOWNER w/ exceptions if field has no |
HAS_EXPECTEDBUCKETOWNER() |
Determine if EXPECTEDBUCKETOWNER has a value |
GlueCatalogGeneration¶
When true, allows Glue to catalog your S3 bucket. Creating an Glue catalog lets you use Athena to query your data.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_GLUECATALOGGENERATION() |
Getter for GLUECATALOGGENERATION, with configurable default |
ASK_GLUECATALOGGENERATION() |
Getter for GLUECATALOGGENERATION w/ exceptions if field has |
HAS_GLUECATALOGGENERATION() |
Determine if GLUECATALOGGENERATION has a value |