Yandex Cloud
Search
Contact UsTry it for free
  • Customer Stories
  • Documentation
  • Blog
  • All Services
  • System Status
  • Marketplace
    • Featured
    • Infrastructure & Network
    • Data Platform
    • AI for business
    • Security
    • DevOps tools
    • Serverless
    • Monitoring & Resources
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Start testing with double trial credits
    • Cloud credits to scale your IT product
    • Gateway to Russia
    • Cloud for Startups
    • Center for Technologies and Society
    • Yandex Cloud Partner program
    • Price calculator
    • Pricing plans
  • Customer Stories
  • Documentation
  • Blog
© 2026 Direct Cursus Technology L.L.C.
Terraform in Yandex Cloud
  • Getting started
  • Solution library
    • Overview
    • Release notes
          • spark_cluster

In this article:

  • Example usage
  • Arguments & Attributes Reference
  1. Terraform reference
  2. Resources
  3. Managed Spark
  4. Data Sources
  5. spark_cluster

yandex_spark_cluster (DataSource)

Written by
Yandex Cloud
Updated at February 20, 2026
  • Example usage
  • Arguments & Attributes Reference

Managed Spark cluster.

Example usageExample usage

//
// Get information about Apache Spark cluster by name
//
data "yandex_spark_cluster" "spark_cluster_by_name" {
  name = "spark-created-with-terraform"
}

//
// Get information about Apache Spark cluster by id
//
data "yandex_spark_cluster" "spark_cluster_by_id" {
  id = "<spark-cluster-id>"
}

Arguments & Attributes ReferenceArguments & Attributes Reference

  • config [Block]. Configuration of the Spark cluster.
    • dependencies [Block]. Environment dependencies.
      • deb_packages (Set Of String). Deb-packages that need to be installed using system package manager.
      • pip_packages (Set Of String). Python packages that need to be installed using pip (in pip requirement format).
    • history_server [Block]. History Server configuration.
      • enabled (Bool). Enable Spark History Server. Default: true.
    • metastore [Block]. Metastore configuration.
      • cluster_id (String). Metastore cluster ID for default spark configuration.
    • resource_pools [Block]. Computational resources.
      • driver [Block]. Computational resources for the driver pool.
        • max_size (Number). Maximum node count for the driver pool with autoscaling.
        • min_size (Number). Minimum node count for the driver pool with autoscaling.
        • resource_preset_id (Required)(String). Resource preset ID for the driver pool.
        • size (Number). Node count for the driver pool with fixed size.
      • executor [Block]. Computational resources for the executor pool.
        • max_size (Number). Maximum node count for the executor pool with autoscaling.
        • min_size (Number). Minimum node count for the executor pool with autoscaling.
        • resource_preset_id (Required)(String). Resource preset ID for the executor pool.
        • size (Number). Node count for the executor pool with fixed size.
    • spark_version (String). Version of Apache Spark.
  • created_at (Read-Only) (String). The timestamp when the cluster was created.
  • deletion_protection (Bool). The true value means that resource is protected from accidental deletion.
  • description (String). Description of the cluster. 0-256 characters long.
  • folder_id (String). ID of the cloud folder that the cluster belongs to.
  • id (Read-Only) (String). Unique ID of the cluster.
  • labels (Map Of String). Cluster labels as key/value pairs.
  • logging [Block]. Cloud Logging configuration.
    • enabled (Bool). Enable log delivery to Cloud Logging. Default: true.
    • folder_id (String). Logs will be written to default log group of specified folder. Exactly one of the attributes folder_id or log_group_id should be specified.
    • log_group_id (String). Logs will be written to the specified log group. Exactly one of the attributes folder_id or log_group_id should be specified.
  • maintenance_window [Block]. Configuration of the window for maintenance operations.
    • day (String). Day of week for maintenance window. One of MON, TUE, WED, THU, FRI, SAT, SUN.
    • hour (Number). Hour of day in UTC time zone (1-24) for maintenance window.
    • type (String). Type of maintenance window. Can be either ANYTIME or WEEKLY. If WEEKLY, day and hour must be specified.
  • name (Required)(String). Name of the cluster. The name is unique within the folder.
  • network [Block]. Network configuration.
    • security_group_ids (Set Of String). Network security groups.
    • subnet_ids (Required)(Set Of String). Network subnets.
  • service_account_id (Required)(String). The service account used by the cluster to access cloud resources.
  • status (Read-Only) (String). Status of the cluster.

Was the article helpful?

Previous
kubernetes_node_group
Next
spark_cluster
© 2026 Direct Cursus Technology L.L.C.