Get Crawler
glue_get_crawler | R Documentation |
Retrieves metadata for a specified crawler¶
Description¶
Retrieves metadata for a specified crawler.
Usage¶
Arguments¶
Name
[required] The name of the crawler to retrieve metadata for.
Value¶
A list with the following syntax:
list(
Crawler = list(
Name = "string",
Role = "string",
Targets = list(
S3Targets = list(
list(
Path = "string",
Exclusions = list(
"string"
),
ConnectionName = "string",
SampleSize = 123,
EventQueueArn = "string",
DlqEventQueueArn = "string"
)
),
JdbcTargets = list(
list(
ConnectionName = "string",
Path = "string",
Exclusions = list(
"string"
),
EnableAdditionalMetadata = list(
"COMMENTS"|"RAWTYPES"
)
)
),
MongoDBTargets = list(
list(
ConnectionName = "string",
Path = "string",
ScanAll = TRUE|FALSE
)
),
DynamoDBTargets = list(
list(
Path = "string",
scanAll = TRUE|FALSE,
scanRate = 123.0
)
),
CatalogTargets = list(
list(
DatabaseName = "string",
Tables = list(
"string"
),
ConnectionName = "string",
EventQueueArn = "string",
DlqEventQueueArn = "string"
)
),
DeltaTargets = list(
list(
DeltaTables = list(
"string"
),
ConnectionName = "string",
WriteManifest = TRUE|FALSE,
CreateNativeDeltaTable = TRUE|FALSE
)
),
IcebergTargets = list(
list(
Paths = list(
"string"
),
ConnectionName = "string",
Exclusions = list(
"string"
),
MaximumTraversalDepth = 123
)
),
HudiTargets = list(
list(
Paths = list(
"string"
),
ConnectionName = "string",
Exclusions = list(
"string"
),
MaximumTraversalDepth = 123
)
)
),
DatabaseName = "string",
Description = "string",
Classifiers = list(
"string"
),
RecrawlPolicy = list(
RecrawlBehavior = "CRAWL_EVERYTHING"|"CRAWL_NEW_FOLDERS_ONLY"|"CRAWL_EVENT_MODE"
),
SchemaChangePolicy = list(
UpdateBehavior = "LOG"|"UPDATE_IN_DATABASE",
DeleteBehavior = "LOG"|"DELETE_FROM_DATABASE"|"DEPRECATE_IN_DATABASE"
),
LineageConfiguration = list(
CrawlerLineageSettings = "ENABLE"|"DISABLE"
),
State = "READY"|"RUNNING"|"STOPPING",
TablePrefix = "string",
Schedule = list(
ScheduleExpression = "string",
State = "SCHEDULED"|"NOT_SCHEDULED"|"TRANSITIONING"
),
CrawlElapsedTime = 123,
CreationTime = as.POSIXct(
"2015-01-01"
),
LastUpdated = as.POSIXct(
"2015-01-01"
),
LastCrawl = list(
Status = "SUCCEEDED"|"CANCELLED"|"FAILED",
ErrorMessage = "string",
LogGroup = "string",
LogStream = "string",
MessagePrefix = "string",
StartTime = as.POSIXct(
"2015-01-01"
)
),
Version = 123,
Configuration = "string",
CrawlerSecurityConfiguration = "string",
LakeFormationConfiguration = list(
UseLakeFormationCredentials = TRUE|FALSE,
AccountId = "string"
)
)
)