1. Packages
  2. AWS Native
  3. API Docs
  4. glue
  5. getCrawler

We recommend new projects start with resources from the AWS provider.

AWS Native v0.126.0 published on Monday, Sep 30, 2024 by Pulumi

aws-native.glue.getCrawler

Explore with Pulumi AI

aws-native logo

We recommend new projects start with resources from the AWS provider.

AWS Native v0.126.0 published on Monday, Sep 30, 2024 by Pulumi

    Resource Type definition for AWS::Glue::Crawler

    Using getCrawler

    Two invocation forms are available. The direct form accepts plain arguments and either blocks until the result value is available, or returns a Promise-wrapped result. The output form accepts Input-wrapped arguments and returns an Output-wrapped result.

    function getCrawler(args: GetCrawlerArgs, opts?: InvokeOptions): Promise<GetCrawlerResult>
    function getCrawlerOutput(args: GetCrawlerOutputArgs, opts?: InvokeOptions): Output<GetCrawlerResult>
    def get_crawler(name: Optional[str] = None,
                    opts: Optional[InvokeOptions] = None) -> GetCrawlerResult
    def get_crawler_output(name: Optional[pulumi.Input[str]] = None,
                    opts: Optional[InvokeOptions] = None) -> Output[GetCrawlerResult]
    func LookupCrawler(ctx *Context, args *LookupCrawlerArgs, opts ...InvokeOption) (*LookupCrawlerResult, error)
    func LookupCrawlerOutput(ctx *Context, args *LookupCrawlerOutputArgs, opts ...InvokeOption) LookupCrawlerResultOutput

    > Note: This function is named LookupCrawler in the Go SDK.

    public static class GetCrawler 
    {
        public static Task<GetCrawlerResult> InvokeAsync(GetCrawlerArgs args, InvokeOptions? opts = null)
        public static Output<GetCrawlerResult> Invoke(GetCrawlerInvokeArgs args, InvokeOptions? opts = null)
    }
    public static CompletableFuture<GetCrawlerResult> getCrawler(GetCrawlerArgs args, InvokeOptions options)
    // Output-based functions aren't available in Java yet
    
    fn::invoke:
      function: aws-native:glue:getCrawler
      arguments:
        # arguments dictionary

    The following arguments are supported:

    Name string
    The name of the crawler.
    Name string
    The name of the crawler.
    name String
    The name of the crawler.
    name string
    The name of the crawler.
    name str
    The name of the crawler.
    name String
    The name of the crawler.

    getCrawler Result

    The following output properties are available:

    Classifiers List<string>
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    Configuration string
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    CrawlerSecurityConfiguration string
    The name of the SecurityConfiguration structure to be used by this crawler.
    DatabaseName string
    The name of the database in which the crawler's output is stored.
    Description string
    A description of the crawler.
    LakeFormationConfiguration Pulumi.AwsNative.Glue.Outputs.CrawlerLakeFormationConfiguration
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    RecrawlPolicy Pulumi.AwsNative.Glue.Outputs.CrawlerRecrawlPolicy
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    Role string
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    Schedule Pulumi.AwsNative.Glue.Outputs.CrawlerSchedule
    For scheduled crawlers, the schedule when the crawler runs.
    SchemaChangePolicy Pulumi.AwsNative.Glue.Outputs.CrawlerSchemaChangePolicy

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    TablePrefix string
    The prefix added to the names of tables that are created.
    Tags object

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    Targets Pulumi.AwsNative.Glue.Outputs.CrawlerTargets
    A collection of targets to crawl.
    Classifiers []string
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    Configuration string
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    CrawlerSecurityConfiguration string
    The name of the SecurityConfiguration structure to be used by this crawler.
    DatabaseName string
    The name of the database in which the crawler's output is stored.
    Description string
    A description of the crawler.
    LakeFormationConfiguration CrawlerLakeFormationConfiguration
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    RecrawlPolicy CrawlerRecrawlPolicy
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    Role string
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    Schedule CrawlerSchedule
    For scheduled crawlers, the schedule when the crawler runs.
    SchemaChangePolicy CrawlerSchemaChangePolicy

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    TablePrefix string
    The prefix added to the names of tables that are created.
    Tags interface{}

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    Targets CrawlerTargets
    A collection of targets to crawl.
    classifiers List<String>
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    configuration String
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    crawlerSecurityConfiguration String
    The name of the SecurityConfiguration structure to be used by this crawler.
    databaseName String
    The name of the database in which the crawler's output is stored.
    description String
    A description of the crawler.
    lakeFormationConfiguration CrawlerLakeFormationConfiguration
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    recrawlPolicy CrawlerRecrawlPolicy
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    role String
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    schedule CrawlerSchedule
    For scheduled crawlers, the schedule when the crawler runs.
    schemaChangePolicy CrawlerSchemaChangePolicy

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    tablePrefix String
    The prefix added to the names of tables that are created.
    tags Object

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    targets CrawlerTargets
    A collection of targets to crawl.
    classifiers string[]
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    configuration string
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    crawlerSecurityConfiguration string
    The name of the SecurityConfiguration structure to be used by this crawler.
    databaseName string
    The name of the database in which the crawler's output is stored.
    description string
    A description of the crawler.
    lakeFormationConfiguration CrawlerLakeFormationConfiguration
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    recrawlPolicy CrawlerRecrawlPolicy
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    role string
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    schedule CrawlerSchedule
    For scheduled crawlers, the schedule when the crawler runs.
    schemaChangePolicy CrawlerSchemaChangePolicy

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    tablePrefix string
    The prefix added to the names of tables that are created.
    tags any

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    targets CrawlerTargets
    A collection of targets to crawl.
    classifiers Sequence[str]
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    configuration str
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    crawler_security_configuration str
    The name of the SecurityConfiguration structure to be used by this crawler.
    database_name str
    The name of the database in which the crawler's output is stored.
    description str
    A description of the crawler.
    lake_formation_configuration CrawlerLakeFormationConfiguration
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    recrawl_policy CrawlerRecrawlPolicy
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    role str
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    schedule CrawlerSchedule
    For scheduled crawlers, the schedule when the crawler runs.
    schema_change_policy CrawlerSchemaChangePolicy

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    table_prefix str
    The prefix added to the names of tables that are created.
    tags Any

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    targets CrawlerTargets
    A collection of targets to crawl.
    classifiers List<String>
    A list of UTF-8 strings that specify the names of custom classifiers that are associated with the crawler.
    configuration String
    Crawler configuration information. This versioned JSON string allows users to specify aspects of a crawler's behavior.
    crawlerSecurityConfiguration String
    The name of the SecurityConfiguration structure to be used by this crawler.
    databaseName String
    The name of the database in which the crawler's output is stored.
    description String
    A description of the crawler.
    lakeFormationConfiguration Property Map
    Specifies whether the crawler should use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    recrawlPolicy Property Map
    A policy that specifies whether to crawl the entire dataset again, or to crawl only folders that were added since the last crawler run.
    role String
    The Amazon Resource Name (ARN) of an IAM role that's used to access customer resources, such as Amazon Simple Storage Service (Amazon S3) data.
    schedule Property Map
    For scheduled crawlers, the schedule when the crawler runs.
    schemaChangePolicy Property Map

    The policy that specifies update and delete behaviors for the crawler. The policy tells the crawler what to do in the event that it detects a change in a table that already exists in the customer's database at the time of the crawl. The SchemaChangePolicy does not affect whether or how new tables and partitions are added. New tables and partitions are always created regardless of the SchemaChangePolicy on a crawler.

    The SchemaChangePolicy consists of two components, UpdateBehavior and DeleteBehavior .

    tablePrefix String
    The prefix added to the names of tables that are created.
    tags Any

    The tags to use with this crawler.

    Search the CloudFormation User Guide for AWS::Glue::Crawler for more information about the expected schema for this property.

    targets Property Map
    A collection of targets to crawl.

    Supporting Types

    CrawlerCatalogTarget

    ConnectionName string
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    DatabaseName string
    The name of the database to be synchronized.
    DlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    EventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    Tables List<string>
    A list of the tables to be synchronized.
    ConnectionName string
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    DatabaseName string
    The name of the database to be synchronized.
    DlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    EventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    Tables []string
    A list of the tables to be synchronized.
    connectionName String
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    databaseName String
    The name of the database to be synchronized.
    dlqEventQueueArn String
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn String
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    tables List<String>
    A list of the tables to be synchronized.
    connectionName string
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    databaseName string
    The name of the database to be synchronized.
    dlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    tables string[]
    A list of the tables to be synchronized.
    connection_name str
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    database_name str
    The name of the database to be synchronized.
    dlq_event_queue_arn str
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    event_queue_arn str
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    tables Sequence[str]
    A list of the tables to be synchronized.
    connectionName String
    The name of the connection for an Amazon S3-backed Data Catalog table to be a target of the crawl when using a Catalog connection type paired with a NETWORK Connection type.
    databaseName String
    The name of the database to be synchronized.
    dlqEventQueueArn String
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn String
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    tables List<String>
    A list of the tables to be synchronized.

    CrawlerDeltaTarget

    ConnectionName string
    The name of the connection to use to connect to the Delta table target.
    CreateNativeDeltaTable bool
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    DeltaTables List<string>
    A list of the Amazon S3 paths to the Delta tables.
    WriteManifest bool
    Specifies whether to write the manifest files to the Delta table path.
    ConnectionName string
    The name of the connection to use to connect to the Delta table target.
    CreateNativeDeltaTable bool
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    DeltaTables []string
    A list of the Amazon S3 paths to the Delta tables.
    WriteManifest bool
    Specifies whether to write the manifest files to the Delta table path.
    connectionName String
    The name of the connection to use to connect to the Delta table target.
    createNativeDeltaTable Boolean
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    deltaTables List<String>
    A list of the Amazon S3 paths to the Delta tables.
    writeManifest Boolean
    Specifies whether to write the manifest files to the Delta table path.
    connectionName string
    The name of the connection to use to connect to the Delta table target.
    createNativeDeltaTable boolean
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    deltaTables string[]
    A list of the Amazon S3 paths to the Delta tables.
    writeManifest boolean
    Specifies whether to write the manifest files to the Delta table path.
    connection_name str
    The name of the connection to use to connect to the Delta table target.
    create_native_delta_table bool
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    delta_tables Sequence[str]
    A list of the Amazon S3 paths to the Delta tables.
    write_manifest bool
    Specifies whether to write the manifest files to the Delta table path.
    connectionName String
    The name of the connection to use to connect to the Delta table target.
    createNativeDeltaTable Boolean
    Specifies whether the crawler will create native tables, to allow integration with query engines that support querying of the Delta transaction log directly.
    deltaTables List<String>
    A list of the Amazon S3 paths to the Delta tables.
    writeManifest Boolean
    Specifies whether to write the manifest files to the Delta table path.

    CrawlerDynamoDbTarget

    Path string
    The name of the DynamoDB table to crawl.
    Path string
    The name of the DynamoDB table to crawl.
    path String
    The name of the DynamoDB table to crawl.
    path string
    The name of the DynamoDB table to crawl.
    path str
    The name of the DynamoDB table to crawl.
    path String
    The name of the DynamoDB table to crawl.

    CrawlerIcebergTarget

    ConnectionName string
    The name of the connection to use to connect to the Iceberg target.
    Exclusions List<string>
    A list of global patterns used to exclude from the crawl.
    MaximumTraversalDepth int
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    Paths List<string>
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
    ConnectionName string
    The name of the connection to use to connect to the Iceberg target.
    Exclusions []string
    A list of global patterns used to exclude from the crawl.
    MaximumTraversalDepth int
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    Paths []string
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
    connectionName String
    The name of the connection to use to connect to the Iceberg target.
    exclusions List<String>
    A list of global patterns used to exclude from the crawl.
    maximumTraversalDepth Integer
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    paths List<String>
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
    connectionName string
    The name of the connection to use to connect to the Iceberg target.
    exclusions string[]
    A list of global patterns used to exclude from the crawl.
    maximumTraversalDepth number
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    paths string[]
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
    connection_name str
    The name of the connection to use to connect to the Iceberg target.
    exclusions Sequence[str]
    A list of global patterns used to exclude from the crawl.
    maximum_traversal_depth int
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    paths Sequence[str]
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .
    connectionName String
    The name of the connection to use to connect to the Iceberg target.
    exclusions List<String>
    A list of global patterns used to exclude from the crawl.
    maximumTraversalDepth Number
    The maximum depth of Amazon S3 paths that the crawler can traverse to discover the Iceberg metadata folder in your Amazon S3 path. Used to limit the crawler run time.
    paths List<String>
    One or more Amazon S3 paths that contains Iceberg metadata folders as s3://bucket/prefix .

    CrawlerJdbcTarget

    ConnectionName string
    The name of the connection to use to connect to the JDBC target.
    EnableAdditionalMetadata List<string>

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    Exclusions List<string>
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    Path string
    The path of the JDBC target.
    ConnectionName string
    The name of the connection to use to connect to the JDBC target.
    EnableAdditionalMetadata []string

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    Exclusions []string
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    Path string
    The path of the JDBC target.
    connectionName String
    The name of the connection to use to connect to the JDBC target.
    enableAdditionalMetadata List<String>

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    exclusions List<String>
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    path String
    The path of the JDBC target.
    connectionName string
    The name of the connection to use to connect to the JDBC target.
    enableAdditionalMetadata string[]

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    exclusions string[]
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    path string
    The path of the JDBC target.
    connection_name str
    The name of the connection to use to connect to the JDBC target.
    enable_additional_metadata Sequence[str]

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    exclusions Sequence[str]
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    path str
    The path of the JDBC target.
    connectionName String
    The name of the connection to use to connect to the JDBC target.
    enableAdditionalMetadata List<String>

    Specify a value of RAWTYPES or COMMENTS to enable additional metadata in table responses. RAWTYPES provides the native-level datatype. COMMENTS provides comments associated with a column or table in the database.

    If you do not need additional metadata, keep the field empty.

    exclusions List<String>
    A list of glob patterns used to exclude from the crawl. For more information, see Catalog Tables with a Crawler.
    path String
    The path of the JDBC target.

    CrawlerLakeFormationConfiguration

    AccountId string
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    UseLakeFormationCredentials bool
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    AccountId string
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    UseLakeFormationCredentials bool
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    accountId String
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    useLakeFormationCredentials Boolean
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    accountId string
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    useLakeFormationCredentials boolean
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    account_id str
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    use_lake_formation_credentials bool
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.
    accountId String
    Required for cross account crawls. For same account crawls as the target data, this can be left as null.
    useLakeFormationCredentials Boolean
    Specifies whether to use AWS Lake Formation credentials for the crawler instead of the IAM role credentials.

    CrawlerMongoDbTarget

    ConnectionName string
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    Path string
    The path of the Amazon DocumentDB or MongoDB target (database/collection).
    ConnectionName string
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    Path string
    The path of the Amazon DocumentDB or MongoDB target (database/collection).
    connectionName String
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    path String
    The path of the Amazon DocumentDB or MongoDB target (database/collection).
    connectionName string
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    path string
    The path of the Amazon DocumentDB or MongoDB target (database/collection).
    connection_name str
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    path str
    The path of the Amazon DocumentDB or MongoDB target (database/collection).
    connectionName String
    The name of the connection to use to connect to the Amazon DocumentDB or MongoDB target.
    path String
    The path of the Amazon DocumentDB or MongoDB target (database/collection).

    CrawlerRecrawlPolicy

    RecrawlBehavior string
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
    RecrawlBehavior string
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
    recrawlBehavior String
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
    recrawlBehavior string
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
    recrawl_behavior str
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.
    recrawlBehavior String
    Specifies whether to crawl the entire dataset again or to crawl only folders that were added since the last crawler run. A value of CRAWL_EVERYTHING specifies crawling the entire dataset again. A value of CRAWL_NEW_FOLDERS_ONLY specifies crawling only folders that were added since the last crawler run. A value of CRAWL_EVENT_MODE specifies crawling only the changes identified by Amazon S3 events.

    CrawlerS3Target

    ConnectionName string
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    DlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    EventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    Exclusions List<string>
    A list of glob patterns used to exclude from the crawl.
    Path string
    The path to the Amazon S3 target.
    SampleSize int
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
    ConnectionName string
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    DlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    EventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    Exclusions []string
    A list of glob patterns used to exclude from the crawl.
    Path string
    The path to the Amazon S3 target.
    SampleSize int
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
    connectionName String
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    dlqEventQueueArn String
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn String
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    exclusions List<String>
    A list of glob patterns used to exclude from the crawl.
    path String
    The path to the Amazon S3 target.
    sampleSize Integer
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
    connectionName string
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    dlqEventQueueArn string
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn string
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    exclusions string[]
    A list of glob patterns used to exclude from the crawl.
    path string
    The path to the Amazon S3 target.
    sampleSize number
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
    connection_name str
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    dlq_event_queue_arn str
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    event_queue_arn str
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    exclusions Sequence[str]
    A list of glob patterns used to exclude from the crawl.
    path str
    The path to the Amazon S3 target.
    sample_size int
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.
    connectionName String
    The name of a connection which allows a job or crawler to access data in Amazon S3 within an Amazon Virtual Private Cloud environment (Amazon VPC).
    dlqEventQueueArn String
    A valid Amazon dead-letter SQS ARN. For example, arn:aws:sqs:region:account:deadLetterQueue.
    eventQueueArn String
    A valid Amazon SQS ARN. For example, arn:aws:sqs:region:account:sqs.
    exclusions List<String>
    A list of glob patterns used to exclude from the crawl.
    path String
    The path to the Amazon S3 target.
    sampleSize Number
    Sets the number of files in each leaf folder to be crawled when crawling sample files in a dataset. If not set, all the files are crawled. A valid value is an integer between 1 and 249.

    CrawlerSchedule

    ScheduleExpression string
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
    ScheduleExpression string
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
    scheduleExpression String
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
    scheduleExpression string
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
    schedule_expression str
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).
    scheduleExpression String
    A cron expression used to specify the schedule. For more information, see Time-Based Schedules for Jobs and Crawlers. For example, to run something every day at 12:15 UTC, specify cron(15 12 * * ? *).

    CrawlerSchemaChangePolicy

    DeleteBehavior string
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    UpdateBehavior string
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
    DeleteBehavior string
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    UpdateBehavior string
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
    deleteBehavior String
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    updateBehavior String
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
    deleteBehavior string
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    updateBehavior string
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
    delete_behavior str
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    update_behavior str
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.
    deleteBehavior String
    The deletion behavior when the crawler finds a deleted object. A value of LOG specifies that if a table or partition is found to no longer exist, do not delete it, only log that it was found to no longer exist. A value of DELETE_FROM_DATABASE specifies that if a table or partition is found to have been removed, delete it from the database. A value of DEPRECATE_IN_DATABASE specifies that if a table has been found to no longer exist, to add a property to the table that says 'DEPRECATED' and includes a timestamp with the time of deprecation.
    updateBehavior String
    The update behavior when the crawler finds a changed schema. A value of LOG specifies that if a table or a partition already exists, and a change is detected, do not update it, only log that a change was detected. Add new tables and new partitions (including on existing tables). A value of UPDATE_IN_DATABASE specifies that if a table or partition already exists, and a change is detected, update it. Add new tables and partitions.

    CrawlerTargets

    CatalogTargets []CrawlerCatalogTarget
    Specifies AWS Glue Data Catalog targets.
    DeltaTargets []CrawlerDeltaTarget
    Specifies an array of Delta data store targets.
    DynamoDbTargets []CrawlerDynamoDbTarget
    Specifies Amazon DynamoDB targets.
    IcebergTargets []CrawlerIcebergTarget
    Specifies Apache Iceberg data store targets.
    JdbcTargets []CrawlerJdbcTarget
    Specifies JDBC targets.
    MongoDbTargets []CrawlerMongoDbTarget
    A list of Mongo DB targets.
    S3Targets []CrawlerS3Target
    Specifies Amazon Simple Storage Service (Amazon S3) targets.
    catalogTargets List<CrawlerCatalogTarget>
    Specifies AWS Glue Data Catalog targets.
    deltaTargets List<CrawlerDeltaTarget>
    Specifies an array of Delta data store targets.
    dynamoDbTargets List<CrawlerDynamoDbTarget>
    Specifies Amazon DynamoDB targets.
    icebergTargets List<CrawlerIcebergTarget>
    Specifies Apache Iceberg data store targets.
    jdbcTargets List<CrawlerJdbcTarget>
    Specifies JDBC targets.
    mongoDbTargets List<CrawlerMongoDbTarget>
    A list of Mongo DB targets.
    s3Targets List<CrawlerS3Target>
    Specifies Amazon Simple Storage Service (Amazon S3) targets.
    catalogTargets CrawlerCatalogTarget[]
    Specifies AWS Glue Data Catalog targets.
    deltaTargets CrawlerDeltaTarget[]
    Specifies an array of Delta data store targets.
    dynamoDbTargets CrawlerDynamoDbTarget[]
    Specifies Amazon DynamoDB targets.
    icebergTargets CrawlerIcebergTarget[]
    Specifies Apache Iceberg data store targets.
    jdbcTargets CrawlerJdbcTarget[]
    Specifies JDBC targets.
    mongoDbTargets CrawlerMongoDbTarget[]
    A list of Mongo DB targets.
    s3Targets CrawlerS3Target[]
    Specifies Amazon Simple Storage Service (Amazon S3) targets.
    catalog_targets Sequence[CrawlerCatalogTarget]
    Specifies AWS Glue Data Catalog targets.
    delta_targets Sequence[CrawlerDeltaTarget]
    Specifies an array of Delta data store targets.
    dynamo_db_targets Sequence[CrawlerDynamoDbTarget]
    Specifies Amazon DynamoDB targets.
    iceberg_targets Sequence[CrawlerIcebergTarget]
    Specifies Apache Iceberg data store targets.
    jdbc_targets Sequence[CrawlerJdbcTarget]
    Specifies JDBC targets.
    mongo_db_targets Sequence[CrawlerMongoDbTarget]
    A list of Mongo DB targets.
    s3_targets Sequence[CrawlerS3Target]
    Specifies Amazon Simple Storage Service (Amazon S3) targets.
    catalogTargets List<Property Map>
    Specifies AWS Glue Data Catalog targets.
    deltaTargets List<Property Map>
    Specifies an array of Delta data store targets.
    dynamoDbTargets List<Property Map>
    Specifies Amazon DynamoDB targets.
    icebergTargets List<Property Map>
    Specifies Apache Iceberg data store targets.
    jdbcTargets List<Property Map>
    Specifies JDBC targets.
    mongoDbTargets List<Property Map>
    A list of Mongo DB targets.
    s3Targets List<Property Map>
    Specifies Amazon Simple Storage Service (Amazon S3) targets.

    Package Details

    Repository
    AWS Native pulumi/pulumi-aws-native
    License
    Apache-2.0
    aws-native logo

    We recommend new projects start with resources from the AWS provider.

    AWS Native v0.126.0 published on Monday, Sep 30, 2024 by Pulumi