Yandex Cloud
Search
Contact UsGet started
  • Blog
  • Pricing
  • Documentation
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • ML & AI
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Customer Stories
    • Gateway to Russia
    • Cloud for Startups
    • Education and Science
  • Blog
  • Pricing
  • Documentation
Yandex project
© 2025 Yandex.Cloud LLC
Yandex Managed Service for Kubernetes
  • Comparison with other Yandex Cloud services
  • Getting started
    • Resource relationships
    • Release channels and updates
    • Encryption
    • Networking in Managed Service for Kubernetes
    • Network settings and cluster policies
    • Automatic scaling
    • Audit policy
    • External cluster nodes
    • Quotas and limits
    • Managed Service for Kubernetes usage recommendations
  • Access management
  • Pricing policy
  • Terraform reference
  • Monitoring metrics
  • Audit Trails events
  • Release notes

In this article:

  • High availability and fault tolerance
  • Load scaling
  • Isolating resources
  1. Concepts
  2. Managed Service for Kubernetes usage recommendations

Tips for using Managed Service for Kubernetes

Written by
Yandex Cloud
Updated at April 18, 2025
  • High availability and fault tolerance
  • Load scaling
  • Isolating resources

Use these recommendations for your PRODUCTION applications that require:

  • High availability and fault tolerance.
  • Load scaling.
  • Resource isolation.

High availability and fault toleranceHigh availability and fault tolerance

  • Use the REGULAR or STABLE release channel. NodeLocal DNS and Pod Topology Constraints are available in all Kubernetes versions.

Tip

Use the RAPID release channel for test environments to more quickly test Kubernetes and Managed Service for Kubernetes updates.

  • Update your cluster and your node groups manually. To do this, disable automatic master and node group updates.

  • Select the highly available master type located in three zones. Kubernetes services will be available in the event of an availability zone level failure. The Managed Service for Kubernetes [Service Level Agreement] applies to the configuration with a highly available master located in three zones.

  • Deploy your Deployment and StatefulSet type services in multiple instances in different availability zones. Use the Pod Topology Constraints and the AntiAffinity strategies to ensure high service availability and efficient usage of Kubernetes cluster resources.

    Use the label combinations below for all strategies:

    • topology.kubernetes.io/zone to keep the services available in the event of an availability zone failure.
    • kubernetes.io/hostname to keep the services available in the event of a cluster node failure.

    Tip

    Test the strategies in a test environment before migrating them to PRODUCTION.

Load scalingLoad scaling

Use these recommendations if the load on your Managed Service for Kubernetes cluster is constantly growing:

  • To make your cluster more robust, create node groups with automatic scaling in multiple availability zones.
  • To reduce the load on the Kubernetes DNS, use NodeLocal DNS. If the cluster contains over 50 nodes, use automatic DNS scaling.
  • To reduce horizontal traffic within the cluster, use a Network Load Balancer and the externalTrafficPolicy:Local rule where possible.
  • Consider node storage requirements in advance:
    • Review disk limits for Yandex Compute Cloud.
    • Load test your disk subsystem in a test environment.
  • To reduce latency at high IOPS, use non-replicated disks.

Isolating resourcesIsolating resources

Follow these recommendations for applications that use shared Kubernetes cluster resources.

Adjust limits and requests for all the cluster services:

---
...
containers:
...
  resources:
    limits:
      cpu: 250m
      memory: 128Mi
    requests:
      cpu: 100m
      memory: 64Mi
...

Specify vCPU availability in thousandths and RAM in megabytes. The service will not exceed the vCPU or RAM limits specified in limits. Customized requests allow you to autoscale cluster nodes.

To manage pod resources automatically, configure Kubernetes policies:

  • Quality of Service for Pods to create pods of different availability classes.
  • Limit Ranges to set limits at the namespace level.

Was the article helpful?

Previous
Quotas and limits
Next
Access management
Yandex project
© 2025 Yandex.Cloud LLC