Skip to content

Get Crawler

glue_get_crawler R Documentation

Retrieves metadata for a specified crawler

Description

Retrieves metadata for a specified crawler.

Usage

glue_get_crawler(Name)

Arguments

Name

[required] The name of the crawler to retrieve metadata for.

Value

A list with the following syntax:

list(
  Crawler = list(
    Name = "string",
    Role = "string",
    Targets = list(
      S3Targets = list(
        list(
          Path = "string",
          Exclusions = list(
            "string"
          ),
          ConnectionName = "string",
          SampleSize = 123,
          EventQueueArn = "string",
          DlqEventQueueArn = "string"
        )
      ),
      JdbcTargets = list(
        list(
          ConnectionName = "string",
          Path = "string",
          Exclusions = list(
            "string"
          ),
          EnableAdditionalMetadata = list(
            "COMMENTS"|"RAWTYPES"
          )
        )
      ),
      MongoDBTargets = list(
        list(
          ConnectionName = "string",
          Path = "string",
          ScanAll = TRUE|FALSE
        )
      ),
      DynamoDBTargets = list(
        list(
          Path = "string",
          scanAll = TRUE|FALSE,
          scanRate = 123.0
        )
      ),
      CatalogTargets = list(
        list(
          DatabaseName = "string",
          Tables = list(
            "string"
          ),
          ConnectionName = "string",
          EventQueueArn = "string",
          DlqEventQueueArn = "string"
        )
      ),
      DeltaTargets = list(
        list(
          DeltaTables = list(
            "string"
          ),
          ConnectionName = "string",
          WriteManifest = TRUE|FALSE,
          CreateNativeDeltaTable = TRUE|FALSE
        )
      ),
      IcebergTargets = list(
        list(
          Paths = list(
            "string"
          ),
          ConnectionName = "string",
          Exclusions = list(
            "string"
          ),
          MaximumTraversalDepth = 123
        )
      ),
      HudiTargets = list(
        list(
          Paths = list(
            "string"
          ),
          ConnectionName = "string",
          Exclusions = list(
            "string"
          ),
          MaximumTraversalDepth = 123
        )
      )
    ),
    DatabaseName = "string",
    Description = "string",
    Classifiers = list(
      "string"
    ),
    RecrawlPolicy = list(
      RecrawlBehavior = "CRAWL_EVERYTHING"|"CRAWL_NEW_FOLDERS_ONLY"|"CRAWL_EVENT_MODE"
    ),
    SchemaChangePolicy = list(
      UpdateBehavior = "LOG"|"UPDATE_IN_DATABASE",
      DeleteBehavior = "LOG"|"DELETE_FROM_DATABASE"|"DEPRECATE_IN_DATABASE"
    ),
    LineageConfiguration = list(
      CrawlerLineageSettings = "ENABLE"|"DISABLE"
    ),
    State = "READY"|"RUNNING"|"STOPPING",
    TablePrefix = "string",
    Schedule = list(
      ScheduleExpression = "string",
      State = "SCHEDULED"|"NOT_SCHEDULED"|"TRANSITIONING"
    ),
    CrawlElapsedTime = 123,
    CreationTime = as.POSIXct(
      "2015-01-01"
    ),
    LastUpdated = as.POSIXct(
      "2015-01-01"
    ),
    LastCrawl = list(
      Status = "SUCCEEDED"|"CANCELLED"|"FAILED",
      ErrorMessage = "string",
      LogGroup = "string",
      LogStream = "string",
      MessagePrefix = "string",
      StartTime = as.POSIXct(
        "2015-01-01"
      )
    ),
    Version = 123,
    Configuration = "string",
    CrawlerSecurityConfiguration = "string",
    LakeFormationConfiguration = list(
      UseLakeFormationCredentials = TRUE|FALSE,
      AccountId = "string"
    )
  )
)

Request syntax

svc$get_crawler(
  Name = "string"
)