Managed Spark API, gRPC: JobService.Get
Returns the specified Spark job.
gRPC request
rpc Get (GetJobRequest) returns (Job)
GetJobRequest
{
"cluster_id": "string",
"job_id": "string"
}
|
Field |
Description |
|
cluster_id |
string Required field. ID of the Spark cluster. |
|
job_id |
string Required field. ID of the Spark job to return. |
Job
{
"id": "string",
"cluster_id": "string",
"created_at": "google.protobuf.Timestamp",
"started_at": "google.protobuf.Timestamp",
"finished_at": "google.protobuf.Timestamp",
"name": "string",
"created_by": "string",
"status": "Status",
// Includes only one of the fields `spark_job`, `pyspark_job`, `spark_connect_job`
"spark_job": {
"args": [
"string"
],
"jar_file_uris": [
"string"
],
"file_uris": [
"string"
],
"archive_uris": [
"string"
],
"properties": "map<string, string>",
"main_jar_file_uri": "string",
"main_class": "string",
"packages": [
"string"
],
"repositories": [
"string"
],
"exclude_packages": [
"string"
]
},
"pyspark_job": {
"args": [
"string"
],
"jar_file_uris": [
"string"
],
"file_uris": [
"string"
],
"archive_uris": [
"string"
],
"properties": "map<string, string>",
"main_python_file_uri": "string",
"python_file_uris": [
"string"
],
"packages": [
"string"
],
"repositories": [
"string"
],
"exclude_packages": [
"string"
]
},
"spark_connect_job": {
"jar_file_uris": [
"string"
],
"file_uris": [
"string"
],
"archive_uris": [
"string"
],
"properties": "map<string, string>",
"packages": [
"string"
],
"repositories": [
"string"
],
"exclude_packages": [
"string"
]
},
// end of the list of possible fields
"ui_url": "string",
"service_account_id": "string",
"connect_url": "string"
}
Spark job.
|
Field |
Description |
|
id |
string Required. Unique ID of the Spark job. |
|
cluster_id |
string Required. Unique ID of the Spark cluster. |
|
created_at |
The time when the Spark job was created. |
|
started_at |
The time when the Spark job was started. |
|
finished_at |
The time when the Spark job was finished. |
|
name |
string Name of the Spark job. |
|
created_by |
string The id of the user who created the job |
|
status |
enum Status Status.
|
|
spark_job |
Includes only one of the fields Job specification. |
|
pyspark_job |
Includes only one of the fields Job specification. |
|
spark_connect_job |
Includes only one of the fields Job specification. |
|
ui_url |
string Spark UI Url. |
|
service_account_id |
string Service account used to access Cloud resources. |
|
connect_url |
string Spark Connect Url. |
SparkJob
|
Field |
Description |
|
args[] |
string Optional arguments to pass to the driver. |
|
jar_file_uris[] |
string Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks. |
|
file_uris[] |
string URIs of files to be copied to the working directory of Spark drivers and distributed tasks. |
|
archive_uris[] |
string URIs of archives to be extracted in the working directory of Spark drivers and tasks. |
|
properties |
object (map<string, string>) A mapping of property names to values, used to configure Spark. |
|
main_jar_file_uri |
string URI of the jar file containing the main class. |
|
main_class |
string The name of the driver's main class. |
|
packages[] |
string List of maven coordinates of jars to include on the driver and executor classpaths. |
|
repositories[] |
string List of additional remote repositories to search for the maven coordinates given with --packages. |
|
exclude_packages[] |
string List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. |
PysparkJob
|
Field |
Description |
|
args[] |
string Optional arguments to pass to the driver. |
|
jar_file_uris[] |
string Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks. |
|
file_uris[] |
string URIs of files to be copied to the working directory of Spark drivers and distributed tasks. |
|
archive_uris[] |
string URIs of archives to be extracted in the working directory of Spark drivers and tasks. |
|
properties |
object (map<string, string>) A mapping of property names to values, used to configure Spark. |
|
main_python_file_uri |
string URI of the main Python file to use as the driver. Must be a .py file. |
|
python_file_uris[] |
string URIs of Python files to pass to the PySpark framework. |
|
packages[] |
string List of maven coordinates of jars to include on the driver and executor classpaths. |
|
repositories[] |
string List of additional remote repositories to search for the maven coordinates given with --packages. |
|
exclude_packages[] |
string List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. |
SparkConnectJob
|
Field |
Description |
|
jar_file_uris[] |
string Jar file URIs to add to the CLASSPATHs of the Spark driver and tasks. |
|
file_uris[] |
string URIs of files to be copied to the working directory of Spark drivers and distributed tasks. |
|
archive_uris[] |
string URIs of archives to be extracted in the working directory of Spark drivers and tasks. |
|
properties |
object (map<string, string>) A mapping of property names to values, used to configure Spark. |
|
packages[] |
string List of maven coordinates of jars to include on the driver and executor classpaths. |
|
repositories[] |
string List of additional remote repositories to search for the maven coordinates given with --packages. |
|
exclude_packages[] |
string List of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts. |