Yandex Cloud
Search
Contact UsTry it for free
  • Customer Stories
  • Documentation
  • Blog
  • All Services
  • System Status
  • Marketplace
    • Featured
    • Infrastructure & Network
    • Data Platform
    • AI for business
    • Security
    • DevOps tools
    • Serverless
    • Monitoring & Resources
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Start testing with double trial credits
    • Cloud credits to scale your IT product
    • Gateway to Russia
    • Cloud for Startups
    • Center for Technologies and Society
    • Yandex Cloud Partner program
    • Price calculator
    • Pricing plans
  • Customer Stories
  • Documentation
  • Blog
© 2026 Direct Cursus Technology L.L.C.
Terraform in Yandex Cloud
  • Getting started
  • Solution library
    • Overview
    • Release notes
          • spark_cluster

In this article:

  • Example usage
  • Arguments & Attributes Reference
  1. Terraform reference
  2. Resources
  3. Managed Spark
  4. Data Sources
  5. spark_cluster

yandex_spark_cluster (DataSource)

Written by
Yandex Cloud
Updated at April 9, 2026
  • Example usage
  • Arguments & Attributes Reference

Managed Spark cluster.

Example usageExample usage

//
// Get information about Apache Spark cluster by name
//
data "yandex_spark_cluster" "spark_cluster_by_name" {
  name = "spark-created-with-terraform"
}

//
// Get information about Apache Spark cluster by id
//
data "yandex_spark_cluster" "spark_cluster_by_id" {
  id = "<spark-cluster-id>"
}

Arguments & Attributes ReferenceArguments & Attributes Reference

  • config [Block]. Configuration of the Spark cluster.
    • dependencies [Block]. Environment dependencies.
      • deb_packages (Set Of String). Deb-packages that need to be installed using system package manager.
      • pip_packages (Set Of String). Python packages that need to be installed using pip (in pip requirement format).
    • history_server [Block]. History Server configuration.
      • enabled (Bool). Enable Spark History Server. Default: true.
    • metastore [Block]. Metastore configuration.
      • cluster_id (String). Metastore cluster ID for default spark configuration.
    • resource_pools [Block]. Computational resources.
      • driver [Block]. Computational resources for the driver pool.
        • max_size (Number). Maximum node count for the driver pool with autoscaling.
        • min_size (Number). Minimum node count for the driver pool with autoscaling.
        • resource_preset_id (Required)(String). Resource preset ID for the driver pool.
        • size (Number). Node count for the driver pool with fixed size.
      • executor [Block]. Computational resources for the executor pool.
        • max_size (Number). Maximum node count for the executor pool with autoscaling.
        • min_size (Number). Minimum node count for the executor pool with autoscaling.
        • resource_preset_id (Required)(String). Resource preset ID for the executor pool.
        • size (Number). Node count for the executor pool with fixed size.
    • spark_version (String). Version of Apache Spark.
  • created_at (Read-Only) (String). The timestamp when the cluster was created.
  • deletion_protection (Bool). The true value means that resource is protected from accidental deletion.
  • description (String). Description of the cluster. 0-256 characters long.
  • folder_id (String). ID of the cloud folder that the cluster belongs to.
  • id (Read-Only) (String). Unique ID of the cluster.
  • labels (Map Of String). Cluster labels as key/value pairs.
  • logging [Block]. Cloud Logging configuration.
    • enabled (Bool). Enable log delivery to Cloud Logging. Default: true.
    • folder_id (String). Logs will be written to default log group of specified folder. Exactly one of the attributes folder_id or log_group_id should be specified.
    • log_group_id (String). Logs will be written to the specified log group. Exactly one of the attributes folder_id or log_group_id should be specified.
  • maintenance_window [Block]. Configuration of the window for maintenance operations.
    • day (String). Day of week for maintenance window. One of MON, TUE, WED, THU, FRI, SAT, SUN.
    • hour (Number). Hour of day in UTC time zone (1-24) for maintenance window.
    • type (String). Type of maintenance window. Can be either ANYTIME or WEEKLY. If WEEKLY, day and hour must be specified.
  • name (Required)(String). Name of the cluster. The name is unique within the folder.
  • network [Block]. Network configuration.
    • security_group_ids (Set Of String). Network security groups.
    • subnet_ids (Required)(Set Of String). Network subnets.
  • service_account_id (Required)(String). The service account used by the cluster to access cloud resources.
  • status (Read-Only) (String). Status of the cluster.
  • timeouts [Block].
    • read (String). A string that can be parsed as a duration consisting of numbers and unit suffixes, such as "30s" or "2h45m". Valid time units are "s" (seconds), "m" (minutes), "h" (hours). Read operations occur during any refresh or planning operation when refresh is enabled.

Was the article helpful?

Previous
kubernetes_node_group
Next
spark_cluster
© 2026 Direct Cursus Technology L.L.C.