Yandex Cloud
Search
Contact UsGet started
  • Blog
  • Pricing
  • Documentation
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • ML & AI
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Customer Stories
    • Gateway to Russia
    • Cloud for Startups
    • Education and Science
  • Blog
  • Pricing
  • Documentation
Yandex project
© 2025 Yandex.Cloud LLC
Yandex Data Processing
  • Getting started
  • Access management
  • Pricing policy
  • Terraform reference
    • Authentication with the API
      • Overview
        • Overview
        • List
        • Create
        • Get
        • ListLog
        • Cancel
  • Monitoring metrics
  • Audit Trails events
  • Public materials
  • FAQ

In this article:

  • HTTP request
  • Path parameters
  • Query parameters
  • Response
  • Job
  • MapreduceJob
  • SparkJob
  • PysparkJob
  • HiveJob
  • QueryList
  • ApplicationInfo
  • ApplicationAttempt
  1. API reference
  2. REST
  3. Job
  4. List

Yandex Data Processing API, REST: Job.List

Written by
Yandex Cloud
Updated at April 2, 2025
  • HTTP request
  • Path parameters
  • Query parameters
  • Response
  • Job
  • MapreduceJob
  • SparkJob
  • PysparkJob
  • HiveJob
  • QueryList
  • ApplicationInfo
  • ApplicationAttempt

Retrieves a list of jobs for a cluster.

HTTP requestHTTP request

GET https://dataproc.api.cloud.yandex.net/dataproc/v1/clusters/{clusterId}/jobs

Path parametersPath parameters

Field

Description

clusterId

string

Required field. ID of the cluster to list jobs for.

Query parametersQuery parameters

Field

Description

pageSize

string (int64)

The maximum number of results per page to return. If the number of available
results is larger than pageSize, the service returns a ListJobsResponse.nextPageToken
that can be used to get the next page of results in subsequent list requests.
Default value: 100.

pageToken

string

Page token. To get the next page of results, set page_token to the
ListJobsResponse.nextPageToken returned by a previous list request.

filter

string

A filter expression that filters jobs listed in the response.

The expression must specify:

  1. The field name. Currently you can use filtering only on Job.name field.
  2. An = operator.
  3. The value in double quotes ("). Must be 3-63 characters long and match the regular expression [a-z][-a-z0-9]{1,61}[a-z0-9].
    Example of a filter: name=my-job.

ResponseResponse

HTTP Code: 200 - OK

{
  "jobs": [
    {
      "id": "string",
      "clusterId": "string",
      "createdAt": "string",
      "startedAt": "string",
      "finishedAt": "string",
      "name": "string",
      "createdBy": "string",
      "status": "string",
      // Includes only one of the fields `mapreduceJob`, `sparkJob`, `pysparkJob`, `hiveJob`
      "mapreduceJob": {
        "args": [
          "string"
        ],
        "jarFileUris": [
          "string"
        ],
        "fileUris": [
          "string"
        ],
        "archiveUris": [
          "string"
        ],
        "properties": "object",
        // Includes only one of the fields `mainJarFileUri`, `mainClass`
        "mainJarFileUri": "string",
        "mainClass": "string"
        // end of the list of possible fields
      },
      "sparkJob": {
        "args": [
          "string"
        ],
        "jarFileUris": [
          "string"
        ],
        "fileUris": [
          "string"
        ],
        "archiveUris": [
          "string"
        ],
        "properties": "object",
        "mainJarFileUri": "string",
        "mainClass": "string",
        "packages": [
          "string"
        ],
        "repositories": [
          "string"
        ],
        "excludePackages": [
          "string"
        ]
      },
      "pysparkJob": {
        "args": [
          "string"
        ],
        "jarFileUris": [
          "string"
        ],
        "fileUris": [
          "string"
        ],
        "archiveUris": [
          "string"
        ],
        "properties": "object",
        "mainPythonFileUri": "string",
        "pythonFileUris": [
          "string"
        ],
        "packages": [
          "string"
        ],
        "repositories": [
          "string"
        ],
        "excludePackages": [
          "string"
        ]
      },
      "hiveJob": {
        "properties": "object",
        "continueOnFailure": "boolean",
        "scriptVariables": "object",
        "jarFileUris": [
          "string"
        ],
        // Includes only one of the fields `queryFileUri`, `queryList`
        "queryFileUri": "string",
        "queryList": {
          "queries": [
            "string"
          ]
        }
        // end of the list of possible fields
      },
      // end of the list of possible fields
      "applicationInfo": {
        "id": "string",
        "applicationAttempts": [
          {
            "id": "string",
            "amContainerId": "string"
          }
        ]
      }
    }
  ],
  "nextPageToken": "string"
}

Field

Description

jobs[]

Job

List of jobs for the specified cluster.

nextPageToken

string

Token for getting the next page of the list. If the number of results is greater than
the specified ListJobsRequest.pageSize, use next_page_token as the value
for the ListJobsRequest.pageToken parameter in the next list request.

Each subsequent page will have its own next_page_token to continue paging through the results.

JobJob

A Yandex Data Processing job. For details about the concept, see documentation.

Field

Description

id

string

ID of the job. Generated at creation time.

clusterId

string

ID of the Yandex Data Processing cluster that the job belongs to.

createdAt

string (date-time)

Creation timestamp.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

startedAt

string (date-time)

The time when the job was started.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

finishedAt

string (date-time)

The time when the job was finished.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

name

string

Name of the job, specified in the JobService.Create request.

createdBy

string

The id of the user who created the job

status

enum (Status)

Job status.

  • STATUS_UNSPECIFIED
  • PROVISIONING: Job is logged in the database and is waiting for the agent to run it.
  • PENDING: Job is acquired by the agent and is in the queue for execution.
  • RUNNING: Job is being run in the cluster.
  • ERROR: Job failed to finish the run properly.
  • DONE: Job is finished.
  • CANCELLED: Job is cancelled.
  • CANCELLING: Job is waiting for cancellation.

mapreduceJob

MapreduceJob

Specification for a MapReduce job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

sparkJob

SparkJob

Specification for a Spark job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

pysparkJob

PysparkJob

Specification for a PySpark job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

hiveJob

HiveJob

Specification for a Hive job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

applicationInfo

ApplicationInfo

Attributes of YARN application.

MapreduceJobMapreduceJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and MapReduce.

mainJarFileUri

string

HCFS URI of the .jar file containing the driver class.

Includes only one of the fields mainJarFileUri, mainClass.

mainClass

string

The name of the driver class.

Includes only one of the fields mainJarFileUri, mainClass.

SparkJobSparkJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and Spark.

mainJarFileUri

string

The HCFS URI of the JAR file containing the main class for the job.

mainClass

string

The name of the driver class.

packages[]

string

List of maven coordinates of jars to include on the driver and executor classpaths.

repositories[]

string

List of additional remote repositories to search for the maven coordinates given with --packages.

excludePackages[]

string

List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts.

PysparkJobPysparkJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and PySpark.

mainPythonFileUri

string

URI of the file with the driver code. Must be a .py file.

pythonFileUris[]

string

URIs of Python files to pass to the PySpark framework.

packages[]

string

List of maven coordinates of jars to include on the driver and executor classpaths.

repositories[]

string

List of additional remote repositories to search for the maven coordinates given with --packages.

excludePackages[]

string

List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts.

HiveJobHiveJob

Field

Description

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and Hive.

continueOnFailure

boolean

Flag indicating whether a job should continue to run if a query fails.

scriptVariables

object (map<string, string>)

Query variables and their values.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Hive driver and each task.

queryFileUri

string

URI of the script with all the necessary Hive queries.

Includes only one of the fields queryFileUri, queryList.

queryList

QueryList

List of Hive queries to be used in the job.

Includes only one of the fields queryFileUri, queryList.

QueryListQueryList

Field

Description

queries[]

string

List of Hive queries.

ApplicationInfoApplicationInfo

Field

Description

id

string

ID of YARN application

applicationAttempts[]

ApplicationAttempt

YARN application attempts

ApplicationAttemptApplicationAttempt

Field

Description

id

string

ID of YARN application attempt

amContainerId

string

ID of YARN Application Master container

Was the article helpful?

Previous
Overview
Next
Create
Yandex project
© 2025 Yandex.Cloud LLC