interface CfnCrawlerMixinProps
| Language | Type name |
|---|---|
.NET | Amazon.CDK.Mixins.Preview.AWS.Glue.Mixins.CfnCrawlerMixinProps |
Go | github.com/aws/aws-cdk-go/awscdkmixinspreview/v2/awsglue/mixins#CfnCrawlerMixinProps |
Java | software.amazon.awscdk.mixins.preview.services.glue.mixins.CfnCrawlerMixinProps |
Python | aws_cdk.mixins_preview.aws_glue.mixins.CfnCrawlerMixinProps |
TypeScript | @aws-cdk/mixins-preview » aws_glue » mixins » CfnCrawlerMixinProps |
Properties for CfnCrawlerPropsMixin.
See also: http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-glue-crawler.html
Example
// The code below shows an example of how to instantiate this type.
// The values are placeholders you should change.
import { mixins as glue_mixins } from '@aws-cdk/mixins-preview/aws-glue';
declare const tags: any;
const cfnCrawlerMixinProps: glue_mixins.CfnCrawlerMixinProps = {
classifiers: ['classifiers'],
configuration: 'configuration',
crawlerSecurityConfiguration: 'crawlerSecurityConfiguration',
databaseName: 'databaseName',
description: 'description',
lakeFormationConfiguration: {
accountId: 'accountId',
useLakeFormationCredentials: false,
},
name: 'name',
recrawlPolicy: {
recrawlBehavior: 'recrawlBehavior',
},
role: 'role',
schedule: {
scheduleExpression: 'scheduleExpression',
},
schemaChangePolicy: {
deleteBehavior: 'deleteBehavior',
updateBehavior: 'updateBehavior',
},
tablePrefix: 'tablePrefix',
tags: tags,
targets: {
catalogTargets: [{
connectionName: 'connectionName',
databaseName: 'databaseName',
dlqEventQueueArn: 'dlqEventQueueArn',
eventQueueArn: 'eventQueueArn',
tables: ['tables'],
}],
deltaTargets: [{
connectionName: 'connectionName',
createNativeDeltaTable: false,
deltaTables: ['deltaTables'],
writeManifest: false,
}],
dynamoDbTargets: [{
path: 'path',
scanAll: false,
scanRate: 123,
}],
hudiTargets: [{
connectionName: 'connectionName',
exclusions: ['exclusions'],
maximumTraversalDepth: 123,
paths: ['paths'],
}],
icebergTargets: [{
connectionName: 'connectionName',
exclusions: ['exclusions'],
maximumTraversalDepth: 123,
paths: ['paths'],
}],
jdbcTargets: [{
connectionName: 'connectionName',
enableAdditionalMetadata: ['enableAdditionalMetadata'],
exclusions: ['exclusions'],
path: 'path',
}],
mongoDbTargets: [{
connectionName: 'connectionName',
path: 'path',
}],
s3Targets: [{
connectionName: 'connectionName',
dlqEventQueueArn: 'dlqEventQueueArn',
eventQueueArn: 'eventQueueArn',
exclusions: ['exclusions'],
path: 'path',
sampleSize: 123,
}],
},
};
Properties
| Name | Type | Description |
|---|---|---|
| classifiers? | string[] | A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler. |
| configuration? | string | Crawler configuration information. |
| crawler | string | The name of the SecurityConfiguration structure to be used by this crawler. |
| database | string | The name of the database in which the crawler's output is stored. |
| description? | string | A description of the crawler. |
| lake | IResolvable | Lake | Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials. |
| name? | string | The name of the crawler. |
| recrawl | IResolvable | Recrawl | A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run. |
| role? | string | The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data. |
| schedule? | IResolvable | Schedule | For scheduled crawlers, the schedule when the crawler runs. |
| schema | IResolvable | Schema | The policy that specifies update and delete behaviors for the crawler. |
| table | string | The prefix added to the names of tables that are created. |
| tags? | any | The tags to use with this crawler. |
| targets? | IResolvable | Targets | A collection of targets to crawl. |
classifiers?
Type:
string[]
(optional)
A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
configuration?
Type:
string
(optional)
Crawler configuration information.
This versioned JSON string allows users to specify aspects of a crawler's behavior. For more information, see Configuring a Crawler .
crawlerSecurityConfiguration?
Type:
string
(optional)
The name of the SecurityConfiguration structure to be used by this crawler.
databaseName?
Type:
string
(optional)
The name of the database in which the crawler's output is stored.
description?
Type:
string
(optional)
A description of the crawler.
lakeFormationConfiguration?
Type:
IResolvable | Lake
(optional)
Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
name?
Type:
string
(optional)
The name of the crawler.
recrawlPolicy?
Type:
IResolvable | Recrawl
(optional)
A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
role?
Type:
string
(optional)
The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
schedule?
Type:
IResolvable | Schedule
(optional)
For scheduled crawlers, the schedule when the crawler runs.
schemaChangePolicy?
Type:
IResolvable | Schema
(optional)
The policy that specifies update and delete behaviors for the crawler.
The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.
The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .
tablePrefix?
Type:
string
(optional)
The prefix added to the names of tables that are created.
tags?
Type:
any
(optional)
The tags to use with this crawler.
targets?
Type:
IResolvable | Targets
(optional)
A collection of targets to crawl.

.NET
Go
Java
Python
TypeScript