Running jobs
Written by
Updated at December 16, 2024
You can run a job in one of the following ways:
- Create jobs in Yandex Data Processing. Once created, they will run automatically.
- Run Apache Hive jobs using the Yandex Cloud CLI or Hive CLI.
- Run Spark or PySpark applications using Spark Shell,
spark-submit
, or the Yandex Cloud CLI. - Use
spark-submit
to run jobs from remote hosts that are not part of the Yandex Data Processing cluster. - Set up integration with Yandex Managed Service for Apache Airflow™ or Yandex DataSphere. This will automate running the jobs.