Yandex Cloud
Search
Contact UsGet started
  • Pricing
  • Customer Stories
  • Documentation
  • Blog
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • AI for business
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Start testing with double trial credits
    • Cloud credits to scale your IT product
    • Gateway to Russia
    • Cloud for Startups
    • Center for Technologies and Society
    • Yandex Cloud Partner program
  • Pricing
  • Customer Stories
  • Documentation
  • Blog
© 2025 Direct Cursus Technology L.L.C.
Yandex Data Transfer
  • Available transfers
  • Getting started
    • All guides
    • Preparing for a transfer
      • Managing endpoints
      • Migrating endpoints to a different availability zone
        • Source
        • Target
    • Managing transfer process
    • Working with databases during transfer
    • Monitoring transfer status
  • Troubleshooting
  • Access management
  • Pricing policy
  • Terraform reference
  • Monitoring metrics
  • Audit Trails events
  • Public materials

In this article:

  • Configuring the data source
  • Preparing the target database
  • Configuring the Elasticsearch target endpoint
  • Custom installation
  • Additional settings
  1. Step-by-step guides
  2. Configuring endpoints
  3. Elasticsearch
  4. Target

Transferring data to an Elasticsearch target endpoint

Written by
Yandex Cloud
Updated at October 24, 2025
  • Configuring the data source
  • Preparing the target database
  • Configuring the Elasticsearch target endpoint
    • Custom installation
    • Additional settings

Yandex Data Transfer enables you to migrate data to a Elasticsearch database and implement various data transfer, processing, and transformation scenarios. To implement a transfer:

  1. Configure one of the supported data sources.
  2. Prepare the Elasticsearch database for the transfer.
  3. Configure the target endpoint in Yandex Data Transfer.
  4. Create a transfer and start it.
  5. Perform required operations with the database and control the transfer.
  6. In case of any issues, use ready-made solutions to resolve them.

Configuring the data sourceConfiguring the data source

Configure one of the supported data sources:

  • Apache Kafka®
  • YDS
  • PostgreSQL
  • Elasticsearch
  • OpenSearch

For a complete list of supported sources and targets in Yandex Data Transfer, see Available transfers.

Preparing the target databasePreparing the target database

Note

Elasticsearch versions lower than 7.x are not supported.

If not planning to use Cloud Interconnect or VPN for connections to an external cluster, make such cluster accessible from the Internet from IP addresses used by Data Transfer.

For details on linking your network up with external resources, see this concept.

  • Make sure the number of columns in the source does not exceed the maximum number of fields in Elasticsearch indexes. The maximum number of fields is provided in the index.mapping.total_fields.limit parameter; the default value is 1000.

    To increase the parameter value, set up a template that makes the maximum number of fields in new indexes equal to the specified value.

    Sample template setup request
    curl \
        --user <Elasticsearch_username>:<password> \
        --header 'Content-Type: application/json' \
        --request PUT "https://<Elasticsearch_cluster_FQDN>:9200/_template/index_defaults" \
        --data '
            {
                "index_patterns": "cdc*",
                "settings": {
                    "index": {
                        "mapping": {
                            "total_fields": {
                                "limit": "2000"
                            }
                        }
                    }
                }
            }'
    

    With this template setup, all new indexes with the cdc* mask may contain up to 2000 fields.

    You can also set up templates using the Kibana interface.

    To check the current index.mapping.total_fields.limit parameter value, use the Kibana interface or execute the following request:

    curl \
        --user <Elasticsearch_username>:<password> \
        --header 'Content-Type: application/json' \
        --request GET 'https://<Elasticsearch_cluster_FQDN>:9200/<index_name>/_settings/*total_fields.limit?include_defaults=true'
    
  • By default, when transferring data to a single index, only one host is used. To distribute the load across hosts when transferring large amounts of data, set up a template to split new indexes into shards in advance.

    Sample template setup request
    curl \
        --user <Elasticsearch_username>:<password> \
        --header 'Content-Type: application/json' \
        --request PUT 'https://<Elasticsearch_cluster_FQDN>:9200/_template/index_defaults' \
        --data '
            {
                "index_patterns": "cdc*",
                "settings" : {
                    "index" : {
                        "number_of_shards" : 15,
                        "number_of_replicas" : 1
                    }
                }
            }'
    

    With this template setup, all new indexes with the cdc* mask will be split into 15 shards.

    You can also set up templates using the Kibana interface.

Configuring the Elasticsearch target endpointConfiguring the Elasticsearch target endpoint

When creating or updating an endpoint, you can define:

  • Elasticsearch connection settings. These are required parameters.
  • Additional parameters.

Custom installationCustom installation

Connection to nodes with explicitly specified network addresses and ports.

Management console
  • Data nodes: Click to add a new data node. For each node, specify:

  • Host: IP address or FQDN of the host with the DATA role you need to connect to.

  • Port: Port number Data Transfer will use to connect to the host with the DATA role.

  • SSL: Select this option if a secure SSL connection is used.

  • CA certificate: Upload the certificate file or add its contents as text if data encryption is required, e.g., to comply with the PCI DSS requirements.

  • Subnet ID: Select or create a subnet in the required availability zone.
    If this field has a value specified for both endpoints, both subnets must be hosted in the same availability zone.

  • User: Specify the username Data Transfer will use to connect to the cluster.

  • Password: Enter the user password to the cluster.

  • Security groups: Select the cloud network to host the endpoint and security groups for network traffic.

    Thus, you will be able to apply the specified security group rules to the VMs and clusters in the selected network without changing the settings of these VMs and clusters. For more information, see Networking in Yandex Data Transfer.

Additional settingsAdditional settings

Management console
  • Cleanup policy: Select a way to clean up data in the target database before the transfer:

    • Don't cleanup: Select this option if you are only going to do replication without copying data.

    • Drop: Completely delete the tables included in the transfer (default).

      Use this option to always transfer the latest version of the table schema to the target database from the source whenever the transfer is activated.

  • Sanitize documents keys: Use this option to automatically replace keys that are not valid for Elasticsearch in the target fields.

    The autocorrect rules are as follows:

    • Empty keys or keys consisting of spaces and periods will be replaced with underscores: "", " ", "." → "_".
    • Leading and trailing periods will be removed: "somekey.", ".somekey" → "somekey".
    • If there are two periods in a row or there is nothing but spaces between them, the entire fragment will be replaced with a period: " some . . key" → " some . key".

    Here is an example of how the autocorrect works: ". s o m e ..incorrect....key. . . " → " s o m e .incorrect.key".

After configuring the data source and target, create and start the transfer.

Was the article helpful?

Previous
Source
Next
Source
© 2025 Direct Cursus Technology L.L.C.