@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class S3Settings extends Object implements Serializable, Cloneable, StructuredPojo
Settings for exporting data to Amazon S3.
| Constructor and Description |
|---|
S3Settings() |
| Modifier and Type | Method and Description |
|---|---|
S3Settings |
clone() |
boolean |
equals(Object obj) |
String |
getBucketFolder()
An optional parameter to set a folder name in the S3 bucket.
|
String |
getBucketName()
The name of the S3 bucket.
|
Boolean |
getCdcInsertsOnly()
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files.
|
String |
getCompressionType()
An optional parameter to use GZIP to compress the target files.
|
String |
getCsvDelimiter()
The delimiter used to separate columns in the source files.
|
String |
getCsvRowDelimiter()
The delimiter used to separate rows in the source files.
|
String |
getDataFormat()
The format of the data that you want to use for output.
|
Integer |
getDataPageSize()
The size of one data page in bytes.
|
Integer |
getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column.
|
Boolean |
getEnableStatistics()
A value that enables statistics for Parquet pages and row groups.
|
String |
getEncodingType()
The type of encoding you are using:
|
String |
getEncryptionMode()
The type of server-side encryption that you want to use for your data.
|
String |
getExternalTableDefinition()
The external table definition.
|
Boolean |
getIncludeOpForFullLoad()
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only
to indicate how the rows were added to the source database.
|
String |
getParquetVersion()
The version of the Apache Parquet format that you want to use:
parquet_1_0 (the default) or
parquet_2_0. |
Integer |
getRowGroupLength()
The number of rows in a row group.
|
String |
getServerSideEncryptionKmsKeyId()
If you are using
SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. |
String |
getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
|
String |
getTimestampColumnName()
A value that includes a timestamp column in the Amazon S3 target endpoint data.
|
int |
hashCode() |
Boolean |
isCdcInsertsOnly()
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files.
|
Boolean |
isEnableStatistics()
A value that enables statistics for Parquet pages and row groups.
|
Boolean |
isIncludeOpForFullLoad()
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only
to indicate how the rows were added to the source database.
|
void |
marshall(ProtocolMarshaller protocolMarshaller) |
void |
setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
void |
setBucketName(String bucketName)
The name of the S3 bucket.
|
void |
setCdcInsertsOnly(Boolean cdcInsertsOnly)
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files.
|
void |
setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
void |
setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
void |
setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
void |
setDataFormat(DataFormatValue dataFormat)
The format of the data that you want to use for output.
|
void |
setDataFormat(String dataFormat)
The format of the data that you want to use for output.
|
void |
setDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
void |
setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
void |
setEnableStatistics(Boolean enableStatistics)
A value that enables statistics for Parquet pages and row groups.
|
void |
setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
|
void |
setEncodingType(String encodingType)
The type of encoding you are using:
|
void |
setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server-side encryption that you want to use for your data.
|
void |
setEncryptionMode(String encryptionMode)
The type of server-side encryption that you want to use for your data.
|
void |
setExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
void |
setIncludeOpForFullLoad(Boolean includeOpForFullLoad)
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only
to indicate how the rows were added to the source database.
|
void |
setParquetVersion(ParquetVersionValue parquetVersion)
The version of the Apache Parquet format that you want to use:
parquet_1_0 (the default) or
parquet_2_0. |
void |
setParquetVersion(String parquetVersion)
The version of the Apache Parquet format that you want to use:
parquet_1_0 (the default) or
parquet_2_0. |
void |
setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
void |
setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using
SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. |
void |
setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
void |
setTimestampColumnName(String timestampColumnName)
A value that includes a timestamp column in the Amazon S3 target endpoint data.
|
String |
toString()
Returns a string representation of this object.
|
S3Settings |
withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket.
|
S3Settings |
withBucketName(String bucketName)
The name of the S3 bucket.
|
S3Settings |
withCdcInsertsOnly(Boolean cdcInsertsOnly)
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files.
|
S3Settings |
withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files.
|
S3Settings |
withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files.
|
S3Settings |
withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files.
|
S3Settings |
withDataFormat(DataFormatValue dataFormat)
The format of the data that you want to use for output.
|
S3Settings |
withDataFormat(String dataFormat)
The format of the data that you want to use for output.
|
S3Settings |
withDataPageSize(Integer dataPageSize)
The size of one data page in bytes.
|
S3Settings |
withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column.
|
S3Settings |
withEnableStatistics(Boolean enableStatistics)
A value that enables statistics for Parquet pages and row groups.
|
S3Settings |
withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
|
S3Settings |
withEncodingType(String encodingType)
The type of encoding you are using:
|
S3Settings |
withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server-side encryption that you want to use for your data.
|
S3Settings |
withEncryptionMode(String encryptionMode)
The type of server-side encryption that you want to use for your data.
|
S3Settings |
withExternalTableDefinition(String externalTableDefinition)
The external table definition.
|
S3Settings |
withIncludeOpForFullLoad(Boolean includeOpForFullLoad)
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only
to indicate how the rows were added to the source database.
|
S3Settings |
withParquetVersion(ParquetVersionValue parquetVersion)
The version of the Apache Parquet format that you want to use:
parquet_1_0 (the default) or
parquet_2_0. |
S3Settings |
withParquetVersion(String parquetVersion)
The version of the Apache Parquet format that you want to use:
parquet_1_0 (the default) or
parquet_2_0. |
S3Settings |
withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group.
|
S3Settings |
withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using
SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. |
S3Settings |
withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
|
S3Settings |
withTimestampColumnName(String timestampColumnName)
A value that includes a timestamp column in the Amazon S3 target endpoint data.
|
public void setServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn - The Amazon Resource Name (ARN) used by the service access IAM role.public String getServiceAccessRoleArn()
The Amazon Resource Name (ARN) used by the service access IAM role.
public S3Settings withServiceAccessRoleArn(String serviceAccessRoleArn)
The Amazon Resource Name (ARN) used by the service access IAM role.
serviceAccessRoleArn - The Amazon Resource Name (ARN) used by the service access IAM role.public void setExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition - The external table definition.public String getExternalTableDefinition()
The external table definition.
public S3Settings withExternalTableDefinition(String externalTableDefinition)
The external table definition.
externalTableDefinition - The external table definition.public void setCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
csvRowDelimiter - The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public String getCsvRowDelimiter()
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
\n).public S3Settings withCsvRowDelimiter(String csvRowDelimiter)
The delimiter used to separate rows in the source files. The default is a carriage return (\n).
csvRowDelimiter - The delimiter used to separate rows in the source files. The default is a carriage return (\n
).public void setCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter - The delimiter used to separate columns in the source files. The default is a comma.public String getCsvDelimiter()
The delimiter used to separate columns in the source files. The default is a comma.
public S3Settings withCsvDelimiter(String csvDelimiter)
The delimiter used to separate columns in the source files. The default is a comma.
csvDelimiter - The delimiter used to separate columns in the source files. The default is a comma.public void setBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter is not specified, then
the path used is schema_name/table_name/.
bucketFolder - An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter is not
specified, then the path used is schema_name/table_name/.public String getBucketFolder()
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter is not specified, then
the path used is schema_name/table_name/.
bucketFolder/schema_name/table_name/. If this parameter is not
specified, then the path used is schema_name/table_name/.public S3Settings withBucketFolder(String bucketFolder)
An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter is not specified, then
the path used is schema_name/table_name/.
bucketFolder - An optional parameter to set a folder name in the S3 bucket. If provided, tables are created in the path
bucketFolder/schema_name/table_name/. If this parameter is not
specified, then the path used is schema_name/table_name/.public void setBucketName(String bucketName)
The name of the S3 bucket.
bucketName - The name of the S3 bucket.public String getBucketName()
The name of the S3 bucket.
public S3Settings withBucketName(String bucketName)
The name of the S3 bucket.
bucketName - The name of the S3 bucket.public void setCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet file formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet
file formats.CompressionTypeValuepublic String getCompressionType()
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet file formats.
CompressionTypeValuepublic S3Settings withCompressionType(String compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet file formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet
file formats.CompressionTypeValuepublic void setCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet file formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet
file formats.CompressionTypeValuepublic S3Settings withCompressionType(CompressionTypeValue compressionType)
An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files. Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet file formats.
compressionType - An optional parameter to use GZIP to compress the target files. Set to GZIP to compress the target files.
Set to NONE (the default) or do not use to leave the files uncompressed. Applies to both .csv and .parquet
file formats.CompressionTypeValuepublic void setEncryptionMode(String encryptionMode)
The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either SSE_S3
(the default) or SSE_KMS. To use SSE_S3, you need an AWS Identity and Access Management
(IAM) role with permission to allow "arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3 (the default) or SSE_KMS. To use SSE_S3, you need an AWS
Identity and Access Management (IAM) role with permission to allow "arn:aws:s3:::dms-*" to
use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic String getEncryptionMode()
The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either SSE_S3
(the default) or SSE_KMS. To use SSE_S3, you need an AWS Identity and Access Management
(IAM) role with permission to allow "arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
SSE_S3 (the default) or SSE_KMS. To use SSE_S3, you need an AWS
Identity and Access Management (IAM) role with permission to allow "arn:aws:s3:::dms-*" to
use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic S3Settings withEncryptionMode(String encryptionMode)
The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either SSE_S3
(the default) or SSE_KMS. To use SSE_S3, you need an AWS Identity and Access Management
(IAM) role with permission to allow "arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3 (the default) or SSE_KMS. To use SSE_S3, you need an AWS
Identity and Access Management (IAM) role with permission to allow "arn:aws:s3:::dms-*" to
use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic void setEncryptionMode(EncryptionModeValue encryptionMode)
The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either SSE_S3
(the default) or SSE_KMS. To use SSE_S3, you need an AWS Identity and Access Management
(IAM) role with permission to allow "arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3 (the default) or SSE_KMS. To use SSE_S3, you need an AWS
Identity and Access Management (IAM) role with permission to allow "arn:aws:s3:::dms-*" to
use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic S3Settings withEncryptionMode(EncryptionModeValue encryptionMode)
The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either SSE_S3
(the default) or SSE_KMS. To use SSE_S3, you need an AWS Identity and Access Management
(IAM) role with permission to allow "arn:aws:s3:::dms-*" to use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
encryptionMode - The type of server-side encryption that you want to use for your data. This encryption type is part of the
endpoint settings or the extra connections attributes for Amazon S3. You can choose either
SSE_S3 (the default) or SSE_KMS. To use SSE_S3, you need an AWS
Identity and Access Management (IAM) role with permission to allow "arn:aws:s3:::dms-*" to
use the following actions:
s3:CreateBucket
s3:ListBucket
s3:DeleteBucket
s3:GetBucketLocation
s3:GetObject
s3:PutObject
s3:DeleteObject
s3:GetObjectVersion
s3:GetBucketPolicy
s3:PutBucketPolicy
s3:DeleteBucketPolicy
EncryptionModeValuepublic void setServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. The key
that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user permissions and
allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
serverSideEncryptionKmsKeyId - If you are using SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. The
key that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user
permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
public String getServerSideEncryptionKmsKeyId()
If you are using SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. The key
that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user permissions and
allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
SSE_KMS for the EncryptionMode, provide the AWS KMS key ID.
The key that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user
permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
public S3Settings withServerSideEncryptionKmsKeyId(String serverSideEncryptionKmsKeyId)
If you are using SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. The key
that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user permissions and
allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
serverSideEncryptionKmsKeyId - If you are using SSE_KMS for the EncryptionMode, provide the AWS KMS key ID. The
key that you use needs an attached policy that enables AWS Identity and Access Management (IAM) user
permissions and allows use of the key.
Here is a CLI example:
aws dms create-endpoint --endpoint-identifier value --endpoint-type target --engine-name s3 --s3-settings ServiceAccessRoleArn=value,BucketFolder=value,BucketName=value,EncryptionMode=SSE_KMS,ServerSideEncryptionKmsKeyId=value
public void setDataFormat(String dataFormat)
The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
dataFormat - The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
DataFormatValuepublic String getDataFormat()
The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features
efficient compression and provides faster query response.
DataFormatValuepublic S3Settings withDataFormat(String dataFormat)
The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
dataFormat - The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
DataFormatValuepublic void setDataFormat(DataFormatValue dataFormat)
The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
dataFormat - The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
DataFormatValuepublic S3Settings withDataFormat(DataFormatValue dataFormat)
The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
dataFormat - The format of the data that you want to use for output. You can choose one of the following:
csv : This is a row-based file format with comma-separated values (.csv).
parquet : Apache Parquet (.parquet) is a columnar storage file format that features efficient
compression and provides faster query response.
DataFormatValuepublic void setEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic String getEncodingType()
The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic S3Settings withEncodingType(String encodingType)
The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic void setEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic S3Settings withEncodingType(EncodingTypeValue encodingType)
The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated values
more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The dictionary is
stored in a dictionary page for each column chunk.
encodingType - The type of encoding you are using:
RLE_DICTIONARY uses a combination of bit-packing and run-length encoding to store repeated
values more efficiently. This is the default.
PLAIN doesn't use encoding at all. Values are stored as they are.
PLAIN_DICTIONARY builds a dictionary of the values encountered in a given column. The
dictionary is stored in a dictionary page for each column chunk.
EncodingTypeValuepublic void setDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. This parameter defaults to 1024 * 1024 bytes (1 MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. This size is used for
.parquet file format only.
dictPageSizeLimit - The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN. This parameter defaults to 1024 * 1024
bytes (1 MiB), the maximum size of a dictionary page before it reverts to PLAIN encoding.
This size is used for .parquet file format only.public Integer getDictPageSizeLimit()
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. This parameter defaults to 1024 * 1024 bytes (1 MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. This size is used for
.parquet file format only.
PLAIN. This parameter defaults to 1024 * 1024
bytes (1 MiB), the maximum size of a dictionary page before it reverts to PLAIN encoding.
This size is used for .parquet file format only.public S3Settings withDictPageSizeLimit(Integer dictPageSizeLimit)
The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this column is
stored using an encoding type of PLAIN. This parameter defaults to 1024 * 1024 bytes (1 MiB), the
maximum size of a dictionary page before it reverts to PLAIN encoding. This size is used for
.parquet file format only.
dictPageSizeLimit - The maximum size of an encoded dictionary page of a column. If the dictionary page exceeds this, this
column is stored using an encoding type of PLAIN. This parameter defaults to 1024 * 1024
bytes (1 MiB), the maximum size of a dictionary page before it reverts to PLAIN encoding.
This size is used for .parquet file format only.public void setRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used for .parquet file format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength - The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used
for .parquet file format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public Integer getRowGroupLength()
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used for .parquet file format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public S3Settings withRowGroupLength(Integer rowGroupLength)
The number of rows in a row group. A smaller row group size provides faster reads. But as the number of row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used for .parquet file format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group length in
bytes (64 * 1024 * 1024).
rowGroupLength - The number of rows in a row group. A smaller row group size provides faster reads. But as the number of
row groups grows, the slower writes become. This parameter defaults to 10,000 rows. This number is used
for .parquet file format only.
If you choose a value larger than the maximum, RowGroupLength is set to the max row group
length in bytes (64 * 1024 * 1024).
public void setDataPageSize(Integer dataPageSize)
The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is used for .parquet file format only.
dataPageSize - The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is
used for .parquet file format only.public Integer getDataPageSize()
The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is used for .parquet file format only.
public S3Settings withDataPageSize(Integer dataPageSize)
The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is used for .parquet file format only.
dataPageSize - The size of one data page in bytes. This parameter defaults to 1024 * 1024 bytes (1 MiB). This number is
used for .parquet file format only.public void setParquetVersion(String parquetVersion)
The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.
parquetVersion - The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.ParquetVersionValuepublic String getParquetVersion()
The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.
parquet_1_0 (the default) or
parquet_2_0.ParquetVersionValuepublic S3Settings withParquetVersion(String parquetVersion)
The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.
parquetVersion - The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.ParquetVersionValuepublic void setParquetVersion(ParquetVersionValue parquetVersion)
The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.
parquetVersion - The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.ParquetVersionValuepublic S3Settings withParquetVersion(ParquetVersionValue parquetVersion)
The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.
parquetVersion - The version of the Apache Parquet format that you want to use: parquet_1_0 (the default) or
parquet_2_0.ParquetVersionValuepublic void setEnableStatistics(Boolean enableStatistics)
A value that enables statistics for Parquet pages and row groups. Choose true to enable statistics,
false to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. This parameter defaults to true. This value is used for .parquet file
format only.
enableStatistics - A value that enables statistics for Parquet pages and row groups. Choose true to enable
statistics, false to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. This parameter defaults to true. This value is
used for .parquet file format only.public Boolean getEnableStatistics()
A value that enables statistics for Parquet pages and row groups. Choose true to enable statistics,
false to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. This parameter defaults to true. This value is used for .parquet file
format only.
true to enable
statistics, false to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. This parameter defaults to true. This value
is used for .parquet file format only.public S3Settings withEnableStatistics(Boolean enableStatistics)
A value that enables statistics for Parquet pages and row groups. Choose true to enable statistics,
false to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. This parameter defaults to true. This value is used for .parquet file
format only.
enableStatistics - A value that enables statistics for Parquet pages and row groups. Choose true to enable
statistics, false to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. This parameter defaults to true. This value is
used for .parquet file format only.public Boolean isEnableStatistics()
A value that enables statistics for Parquet pages and row groups. Choose true to enable statistics,
false to disable. Statistics include NULL, DISTINCT, MAX, and
MIN values. This parameter defaults to true. This value is used for .parquet file
format only.
true to enable
statistics, false to disable. Statistics include NULL, DISTINCT,
MAX, and MIN values. This parameter defaults to true. This value
is used for .parquet file format only.public void setIncludeOpForFullLoad(Boolean includeOpForFullLoad)
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information is
recorded in these output files for a full load to indicate that the rows were inserted at the source database. If
IncludeOpForFullLoad is set to true or y, the INSERT is recorded as an I
annotation in the first field of the .csv file. This allows the format of your target records from a full load to
be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more information
about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
includeOpForFullLoad - A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output
files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information
is recorded in these output files for a full load to indicate that the rows were inserted at the source
database. If IncludeOpForFullLoad is set to true or y, the INSERT
is recorded as an I annotation in the first field of the .csv file. This allows the format of your target
records from a full load to be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
public Boolean getIncludeOpForFullLoad()
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information is
recorded in these output files for a full load to indicate that the rows were inserted at the source database. If
IncludeOpForFullLoad is set to true or y, the INSERT is recorded as an I
annotation in the first field of the .csv file. This allows the format of your target records from a full load to
be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more information
about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information
is recorded in these output files for a full load to indicate that the rows were inserted at the source
database. If IncludeOpForFullLoad is set to true or y, the INSERT
is recorded as an I annotation in the first field of the .csv file. This allows the format of your target
records from a full load to be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
public S3Settings withIncludeOpForFullLoad(Boolean includeOpForFullLoad)
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information is
recorded in these output files for a full load to indicate that the rows were inserted at the source database. If
IncludeOpForFullLoad is set to true or y, the INSERT is recorded as an I
annotation in the first field of the .csv file. This allows the format of your target records from a full load to
be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more information
about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
includeOpForFullLoad - A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output
files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information
is recorded in these output files for a full load to indicate that the rows were inserted at the source
database. If IncludeOpForFullLoad is set to true or y, the INSERT
is recorded as an I annotation in the first field of the .csv file. This allows the format of your target
records from a full load to be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
public Boolean isIncludeOpForFullLoad()
A value that enables a full load to write INSERT operations to the comma-separated value (.csv) output files only to indicate how the rows were added to the source database.
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information is
recorded in these output files for a full load to indicate that the rows were inserted at the source database. If
IncludeOpForFullLoad is set to true or y, the INSERT is recorded as an I
annotation in the first field of the .csv file. This allows the format of your target records from a full load to
be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more information
about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports IncludeOpForFullLoad in versions 3.1.4 and later.
For full load, records can only be inserted. By default (the false setting), no information
is recorded in these output files for a full load to indicate that the rows were inserted at the source
database. If IncludeOpForFullLoad is set to true or y, the INSERT
is recorded as an I annotation in the first field of the .csv file. This allows the format of your target
records from a full load to be consistent with the target records from a CDC load.
This setting works together with CdcInsertsOnly for output to .csv files only. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
public void setCdcInsertsOnly(Boolean cdcInsertsOnly)
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files. By default (the false setting), the first field in a .csv or .parquet
record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was
inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends
on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written without a
first field to indicate the INSERT operation at the source. For more information about how these settings work
together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and IncludeOpForFullLoad in
versions 3.1.4 and later.
cdcInsertsOnly - A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar
storage (.parquet) output files. By default (the false setting), the first field in a .csv or
.parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether
the row was inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded
depends on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written
without a first field to indicate the INSERT operation at the source. For more information about how these
settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and
IncludeOpForFullLoad in versions 3.1.4 and later.
public Boolean getCdcInsertsOnly()
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files. By default (the false setting), the first field in a .csv or .parquet
record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was
inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends
on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written without a
first field to indicate the INSERT operation at the source. For more information about how these settings work
together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and IncludeOpForFullLoad in
versions 3.1.4 and later.
false setting), the first field in a .csv
or .parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate
whether the row was inserted, updated, or deleted at the source database for a CDC load to the
target.
If cdcInsertsOnly is set to true or y, only INSERTs from the
source database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are
recorded depends on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad
is set to true, the first field of every CDC record is set to I to indicate the INSERT
operation at the source. If IncludeOpForFullLoad is set to false, every CDC
record is written without a first field to indicate the INSERT operation at the source. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and
IncludeOpForFullLoad in versions 3.1.4 and later.
public S3Settings withCdcInsertsOnly(Boolean cdcInsertsOnly)
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files. By default (the false setting), the first field in a .csv or .parquet
record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was
inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends
on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written without a
first field to indicate the INSERT operation at the source. For more information about how these settings work
together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and IncludeOpForFullLoad in
versions 3.1.4 and later.
cdcInsertsOnly - A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar
storage (.parquet) output files. By default (the false setting), the first field in a .csv or
.parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether
the row was inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded
depends on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written
without a first field to indicate the INSERT operation at the source. For more information about how these
settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and
IncludeOpForFullLoad in versions 3.1.4 and later.
public Boolean isCdcInsertsOnly()
A value that enables a change data capture (CDC) load to write only INSERT operations to .csv or columnar storage
(.parquet) output files. By default (the false setting), the first field in a .csv or .parquet
record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate whether the row was
inserted, updated, or deleted at the source database for a CDC load to the target.
If cdcInsertsOnly is set to true or y, only INSERTs from the source
database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are recorded depends
on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad is set to
true, the first field of every CDC record is set to I to indicate the INSERT operation at the
source. If IncludeOpForFullLoad is set to false, every CDC record is written without a
first field to indicate the INSERT operation at the source. For more information about how these settings work
together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and IncludeOpForFullLoad in
versions 3.1.4 and later.
false setting), the first field in a .csv
or .parquet record contains the letter I (INSERT), U (UPDATE), or D (DELETE). These values indicate
whether the row was inserted, updated, or deleted at the source database for a CDC load to the
target.
If cdcInsertsOnly is set to true or y, only INSERTs from the
source database are migrated to the .csv or .parquet file. For .csv format only, how these INSERTs are
recorded depends on the value of IncludeOpForFullLoad. If IncludeOpForFullLoad
is set to true, the first field of every CDC record is set to I to indicate the INSERT
operation at the source. If IncludeOpForFullLoad is set to false, every CDC
record is written without a first field to indicate the INSERT operation at the source. For more
information about how these settings work together, see Indicating Source DB Operations in Migrated S3 Data in the AWS Database Migration Service User
Guide..
AWS DMS supports this interaction between CdcInsertsOnly and
IncludeOpForFullLoad in versions 3.1.4 and later.
public void setTimestampColumnName(String timestampColumnName)
A value that includes a timestamp column in the Amazon S3 target endpoint data. AWS DMS includes an additional
column in the migrated data when you set timestampColumnName to a non-blank value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred from the
source to the target by DMS. For a CDC load, each row of the timestamp column contains the timestamp for the
commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit timestamp
supported by DMS for the source database. When the AddColumnName setting is set to true
, DMS also includes the name for the timestamp column that you set as the nonblank value of
timestampColumnName.
timestampColumnName - A value that includes a timestamp column in the Amazon S3 target endpoint data. AWS DMS includes an
additional column in the migrated data when you set timestampColumnName to a non-blank value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred
from the source to the target by DMS. For a CDC load, each row of the timestamp column contains the
timestamp for the commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit
timestamp supported by DMS for the source database. When the AddColumnName setting is set to
true, DMS also includes the name for the timestamp column that you set as the nonblank value
of timestampColumnName.
public String getTimestampColumnName()
A value that includes a timestamp column in the Amazon S3 target endpoint data. AWS DMS includes an additional
column in the migrated data when you set timestampColumnName to a non-blank value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred from the
source to the target by DMS. For a CDC load, each row of the timestamp column contains the timestamp for the
commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit timestamp
supported by DMS for the source database. When the AddColumnName setting is set to true
, DMS also includes the name for the timestamp column that you set as the nonblank value of
timestampColumnName.
timestampColumnName to a non-blank
value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred
from the source to the target by DMS. For a CDC load, each row of the timestamp column contains the
timestamp for the commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit
timestamp supported by DMS for the source database. When the AddColumnName setting is set to
true, DMS also includes the name for the timestamp column that you set as the nonblank value
of timestampColumnName.
public S3Settings withTimestampColumnName(String timestampColumnName)
A value that includes a timestamp column in the Amazon S3 target endpoint data. AWS DMS includes an additional
column in the migrated data when you set timestampColumnName to a non-blank value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred from the
source to the target by DMS. For a CDC load, each row of the timestamp column contains the timestamp for the
commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit timestamp
supported by DMS for the source database. When the AddColumnName setting is set to true
, DMS also includes the name for the timestamp column that you set as the nonblank value of
timestampColumnName.
timestampColumnName - A value that includes a timestamp column in the Amazon S3 target endpoint data. AWS DMS includes an
additional column in the migrated data when you set timestampColumnName to a non-blank value.
AWS DMS supports TimestampColumnName in versions 3.1.4 and later.
For a full load, each row of the timestamp column contains a timestamp for when the data was transferred
from the source to the target by DMS. For a CDC load, each row of the timestamp column contains the
timestamp for the commit of that row in the source database. The format for the timestamp column value is
yyyy-MM-dd HH:mm:ss.SSSSSS. For CDC, the microsecond precision depends on the commit
timestamp supported by DMS for the source database. When the AddColumnName setting is set to
true, DMS also includes the name for the timestamp column that you set as the nonblank value
of timestampColumnName.
public String toString()
toString in class ObjectObject.toString()public S3Settings clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
marshall in interface StructuredPojoCopyright © 2019. All rights reserved.