Skip to content

List Job Runs

emrcontainers_list_job_runs R Documentation

Lists job runs based on a set of parameters

Description

Lists job runs based on a set of parameters. A job run is a unit of work, such as a Spark jar, PySpark script, or SparkSQL query, that you submit to Amazon EMR on EKS.

Usage

emrcontainers_list_job_runs(virtualClusterId, createdBefore,
  createdAfter, name, states, maxResults, nextToken)

Arguments

virtualClusterId

[required] The ID of the virtual cluster for which to list the job run.

createdBefore

The date and time before which the job runs were submitted.

createdAfter

The date and time after which the job runs were submitted.

name

The name of the job run.

states

The states of the job run.

maxResults

The maximum number of job runs that can be listed.

nextToken

The token for the next set of job runs to return.

Value

A list with the following syntax:

list(
  jobRuns = list(
    list(
      id = "string",
      name = "string",
      virtualClusterId = "string",
      arn = "string",
      state = "PENDING"|"SUBMITTED"|"RUNNING"|"FAILED"|"CANCELLED"|"CANCEL_PENDING"|"COMPLETED",
      clientToken = "string",
      executionRoleArn = "string",
      releaseLabel = "string",
      configurationOverrides = list(
        applicationConfiguration = list(
          list(
            classification = "string",
            properties = list(
              "string"
            ),
            configurations = list()
          )
        ),
        monitoringConfiguration = list(
          persistentAppUI = "ENABLED"|"DISABLED",
          cloudWatchMonitoringConfiguration = list(
            logGroupName = "string",
            logStreamNamePrefix = "string"
          ),
          s3MonitoringConfiguration = list(
            logUri = "string"
          ),
          containerLogRotationConfiguration = list(
            rotationSize = "string",
            maxFilesToKeep = 123
          )
        )
      ),
      jobDriver = list(
        sparkSubmitJobDriver = list(
          entryPoint = "string",
          entryPointArguments = list(
            "string"
          ),
          sparkSubmitParameters = "string"
        ),
        sparkSqlJobDriver = list(
          entryPoint = "string",
          sparkSqlParameters = "string"
        )
      ),
      createdAt = as.POSIXct(
        "2015-01-01"
      ),
      createdBy = "string",
      finishedAt = as.POSIXct(
        "2015-01-01"
      ),
      stateDetails = "string",
      failureReason = "INTERNAL_ERROR"|"USER_ERROR"|"VALIDATION_ERROR"|"CLUSTER_UNAVAILABLE",
      tags = list(
        "string"
      ),
      retryPolicyConfiguration = list(
        maxAttempts = 123
      ),
      retryPolicyExecution = list(
        currentAttemptCount = 123
      )
    )
  ),
  nextToken = "string"
)

Request syntax

svc$list_job_runs(
  virtualClusterId = "string",
  createdBefore = as.POSIXct(
    "2015-01-01"
  ),
  createdAfter = as.POSIXct(
    "2015-01-01"
  ),
  name = "string",
  states = list(
    "PENDING"|"SUBMITTED"|"RUNNING"|"FAILED"|"CANCELLED"|"CANCEL_PENDING"|"COMPLETED"
  ),
  maxResults = 123,
  nextToken = "string"
)