Update Recipe Job
gluedatabrew_update_recipe_job | R Documentation |
Modifies the definition of an existing DataBrew recipe job¶
Description¶
Modifies the definition of an existing DataBrew recipe job.
Usage¶
gluedatabrew_update_recipe_job(EncryptionKeyArn, EncryptionMode, Name,
LogSubscription, MaxCapacity, MaxRetries, Outputs, DataCatalogOutputs,
DatabaseOutputs, RoleArn, Timeout)
Arguments¶
EncryptionKeyArn |
The Amazon Resource Name (ARN) of an encryption key that is used to protect the job. |
EncryptionMode |
The encryption mode for the job, which can be one of the following:
|
Name |
[required] The name of the job to update. |
LogSubscription |
Enables or disables Amazon CloudWatch logging for the job. If logging is enabled, CloudWatch writes one log stream for each job run. |
MaxCapacity |
The maximum number of nodes that DataBrew can consume when the job processes data. |
MaxRetries |
The maximum number of times to retry the job after a job run fails. |
Outputs |
One or more artifacts that represent the output from running the job. |
DataCatalogOutputs |
One or more artifacts that represent the Glue Data Catalog output from running the job. |
DatabaseOutputs |
Represents a list of JDBC database output objects which defines the output destination for a DataBrew recipe job to write into. |
RoleArn |
[required] The Amazon Resource Name (ARN) of the Identity and Access Management (IAM) role to be assumed when DataBrew runs the job. |
Timeout |
The job's timeout in minutes. A job that attempts to run longer
than this timeout period ends with a status of
|
Value¶
A list with the following syntax:
list(
Name = "string"
)
Request syntax¶
svc$update_recipe_job(
EncryptionKeyArn = "string",
EncryptionMode = "SSE-KMS"|"SSE-S3",
Name = "string",
LogSubscription = "ENABLE"|"DISABLE",
MaxCapacity = 123,
MaxRetries = 123,
Outputs = list(
list(
CompressionFormat = "GZIP"|"LZ4"|"SNAPPY"|"BZIP2"|"DEFLATE"|"LZO"|"BROTLI"|"ZSTD"|"ZLIB",
Format = "CSV"|"JSON"|"PARQUET"|"GLUEPARQUET"|"AVRO"|"ORC"|"XML"|"TABLEAUHYPER",
PartitionColumns = list(
"string"
),
Location = list(
Bucket = "string",
Key = "string",
BucketOwner = "string"
),
Overwrite = TRUE|FALSE,
FormatOptions = list(
Csv = list(
Delimiter = "string"
)
),
MaxOutputFiles = 123
)
),
DataCatalogOutputs = list(
list(
CatalogId = "string",
DatabaseName = "string",
TableName = "string",
S3Options = list(
Location = list(
Bucket = "string",
Key = "string",
BucketOwner = "string"
)
),
DatabaseOptions = list(
TempDirectory = list(
Bucket = "string",
Key = "string",
BucketOwner = "string"
),
TableName = "string"
),
Overwrite = TRUE|FALSE
)
),
DatabaseOutputs = list(
list(
GlueConnectionName = "string",
DatabaseOptions = list(
TempDirectory = list(
Bucket = "string",
Key = "string",
BucketOwner = "string"
),
TableName = "string"
),
DatabaseOutputMode = "NEW_TABLE"
)
),
RoleArn = "string",
Timeout = 123
)