Yandex Cloud
Search
Contact UsGet started
  • Blog
  • Pricing
  • Documentation
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • ML & AI
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Customer Stories
    • Gateway to Russia
    • Cloud for Startups
    • Education and Science
  • Blog
  • Pricing
  • Documentation
Yandex project
© 2025 Yandex.Cloud LLC
Yandex Data Processing
  • Getting started
  • Access management
  • Pricing policy
  • Terraform reference
    • Authentication with the API
      • Overview
        • Overview
        • List
        • Create
        • Get
        • ListLog
        • Cancel
  • Monitoring metrics
  • Audit Trails events
  • Public materials
  • FAQ

In this article:

  • HTTP request
  • Path parameters
  • Response
  • MapreduceJob
  • SparkJob
  • PysparkJob
  • HiveJob
  • QueryList
  • ApplicationInfo
  • ApplicationAttempt
  1. API reference
  2. REST
  3. Job
  4. Get

Yandex Data Processing API, REST: Job.Get

Written by
Yandex Cloud
Updated at April 2, 2025
  • HTTP request
  • Path parameters
  • Response
  • MapreduceJob
  • SparkJob
  • PysparkJob
  • HiveJob
  • QueryList
  • ApplicationInfo
  • ApplicationAttempt

Returns the specified job.

HTTP requestHTTP request

GET https://dataproc.api.cloud.yandex.net/dataproc/v1/clusters/{clusterId}/jobs/{jobId}

Path parametersPath parameters

Field

Description

clusterId

string

Required field. ID of the cluster to request a job from.

jobId

string

Required field. ID of the job to return.

To get a job ID make a JobService.List request.

ResponseResponse

HTTP Code: 200 - OK

{
  "id": "string",
  "clusterId": "string",
  "createdAt": "string",
  "startedAt": "string",
  "finishedAt": "string",
  "name": "string",
  "createdBy": "string",
  "status": "string",
  // Includes only one of the fields `mapreduceJob`, `sparkJob`, `pysparkJob`, `hiveJob`
  "mapreduceJob": {
    "args": [
      "string"
    ],
    "jarFileUris": [
      "string"
    ],
    "fileUris": [
      "string"
    ],
    "archiveUris": [
      "string"
    ],
    "properties": "object",
    // Includes only one of the fields `mainJarFileUri`, `mainClass`
    "mainJarFileUri": "string",
    "mainClass": "string"
    // end of the list of possible fields
  },
  "sparkJob": {
    "args": [
      "string"
    ],
    "jarFileUris": [
      "string"
    ],
    "fileUris": [
      "string"
    ],
    "archiveUris": [
      "string"
    ],
    "properties": "object",
    "mainJarFileUri": "string",
    "mainClass": "string",
    "packages": [
      "string"
    ],
    "repositories": [
      "string"
    ],
    "excludePackages": [
      "string"
    ]
  },
  "pysparkJob": {
    "args": [
      "string"
    ],
    "jarFileUris": [
      "string"
    ],
    "fileUris": [
      "string"
    ],
    "archiveUris": [
      "string"
    ],
    "properties": "object",
    "mainPythonFileUri": "string",
    "pythonFileUris": [
      "string"
    ],
    "packages": [
      "string"
    ],
    "repositories": [
      "string"
    ],
    "excludePackages": [
      "string"
    ]
  },
  "hiveJob": {
    "properties": "object",
    "continueOnFailure": "boolean",
    "scriptVariables": "object",
    "jarFileUris": [
      "string"
    ],
    // Includes only one of the fields `queryFileUri`, `queryList`
    "queryFileUri": "string",
    "queryList": {
      "queries": [
        "string"
      ]
    }
    // end of the list of possible fields
  },
  // end of the list of possible fields
  "applicationInfo": {
    "id": "string",
    "applicationAttempts": [
      {
        "id": "string",
        "amContainerId": "string"
      }
    ]
  }
}

A Yandex Data Processing job. For details about the concept, see documentation.

Field

Description

id

string

ID of the job. Generated at creation time.

clusterId

string

ID of the Yandex Data Processing cluster that the job belongs to.

createdAt

string (date-time)

Creation timestamp.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

startedAt

string (date-time)

The time when the job was started.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

finishedAt

string (date-time)

The time when the job was finished.

String in RFC3339 text format. The range of possible values is from
0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z, i.e. from 0 to 9 digits for fractions of a second.

To work with values in this field, use the APIs described in the
Protocol Buffers reference.
In some languages, built-in datetime utilities do not support nanosecond precision (9 digits).

name

string

Name of the job, specified in the JobService.Create request.

createdBy

string

The id of the user who created the job

status

enum (Status)

Job status.

  • STATUS_UNSPECIFIED
  • PROVISIONING: Job is logged in the database and is waiting for the agent to run it.
  • PENDING: Job is acquired by the agent and is in the queue for execution.
  • RUNNING: Job is being run in the cluster.
  • ERROR: Job failed to finish the run properly.
  • DONE: Job is finished.
  • CANCELLED: Job is cancelled.
  • CANCELLING: Job is waiting for cancellation.

mapreduceJob

MapreduceJob

Specification for a MapReduce job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

sparkJob

SparkJob

Specification for a Spark job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

pysparkJob

PysparkJob

Specification for a PySpark job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

hiveJob

HiveJob

Specification for a Hive job.

Includes only one of the fields mapreduceJob, sparkJob, pysparkJob, hiveJob.

Specification for the job.

applicationInfo

ApplicationInfo

Attributes of YARN application.

MapreduceJobMapreduceJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and MapReduce.

mainJarFileUri

string

HCFS URI of the .jar file containing the driver class.

Includes only one of the fields mainJarFileUri, mainClass.

mainClass

string

The name of the driver class.

Includes only one of the fields mainJarFileUri, mainClass.

SparkJobSparkJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and Spark.

mainJarFileUri

string

The HCFS URI of the JAR file containing the main class for the job.

mainClass

string

The name of the driver class.

packages[]

string

List of maven coordinates of jars to include on the driver and executor classpaths.

repositories[]

string

List of additional remote repositories to search for the maven coordinates given with --packages.

excludePackages[]

string

List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts.

PysparkJobPysparkJob

Field

Description

args[]

string

Optional arguments to pass to the driver.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Yandex Data Processing driver and each task.

fileUris[]

string

URIs of resource files to be copied to the working directory of Yandex Data Processing drivers
and distributed Hadoop tasks.

archiveUris[]

string

URIs of archives to be extracted to the working directory of Yandex Data Processing drivers and tasks.

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and PySpark.

mainPythonFileUri

string

URI of the file with the driver code. Must be a .py file.

pythonFileUris[]

string

URIs of Python files to pass to the PySpark framework.

packages[]

string

List of maven coordinates of jars to include on the driver and executor classpaths.

repositories[]

string

List of additional remote repositories to search for the maven coordinates given with --packages.

excludePackages[]

string

List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts.

HiveJobHiveJob

Field

Description

properties

object (map<string, string>)

Property names and values, used to configure Yandex Data Processing and Hive.

continueOnFailure

boolean

Flag indicating whether a job should continue to run if a query fails.

scriptVariables

object (map<string, string>)

Query variables and their values.

jarFileUris[]

string

JAR file URIs to add to CLASSPATH of the Hive driver and each task.

queryFileUri

string

URI of the script with all the necessary Hive queries.

Includes only one of the fields queryFileUri, queryList.

queryList

QueryList

List of Hive queries to be used in the job.

Includes only one of the fields queryFileUri, queryList.

QueryListQueryList

Field

Description

queries[]

string

List of Hive queries.

ApplicationInfoApplicationInfo

Field

Description

id

string

ID of YARN application

applicationAttempts[]

ApplicationAttempt

YARN application attempts

ApplicationAttemptApplicationAttempt

Field

Description

id

string

ID of YARN application attempt

amContainerId

string

ID of YARN Application Master container

Was the article helpful?

Previous
Create
Next
ListLog
Yandex project
© 2025 Yandex.Cloud LLC