Yandex Cloud
Search
Contact UsGet started
  • Blog
  • Pricing
  • Documentation
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • ML & AI
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Customer Stories
    • Gateway to Russia
    • Cloud for Startups
    • Education and Science
  • Blog
  • Pricing
  • Documentation
Yandex project
© 2025 Yandex.Cloud LLC
Yandex Managed Service for Apache Airflow™
  • Getting started
    • All guides
      • Uploading a file to Yandex Object Storage
      • Connecting to a Yandex Object Storage bucket with a bucket policy
      • Reading a file from Yandex Object Storage
    • Working with Apache Airflow™ interfaces
    • Transferring logs from Apache Airflow™ to Cloud Logging
  • Access management
  • Pricing policy
  • Terraform reference
  • Yandex Monitoring metrics
  • Release notes
  • FAQ

In this article:

  • Prepare the DAG file and run the graph
  • Check the result
  1. Step-by-step guides
  2. Working with Yandex Object Storage
  3. Reading a file from Yandex Object Storage

Reading a file from Yandex Object Storage

Written by
Yandex Cloud
Updated at April 10, 2025
  • Prepare the DAG file and run the graph
  • Check the result

Use a directed acyclic graph (DAG) to read files from Yandex Object Storage of the service account linked to an Apache Airflow™ cluster.

Prepare the DAG file and run the graphPrepare the DAG file and run the graph

  1. Create a local file named read_file_from_dags_bucket.py and copy the following script to it:

    from airflow.settings import DAGS_FOLDER
    from airflow.decorators import dag, task
    
    
    @dag(schedule=None)
    def read_file_from_dags_bucket():
        @task
        def read_file():
            with open(f'{DAGS_FOLDER}/data/airflow.txt') as file:
                content = file.read()
                print(f"file content: {content}")
    
        read_file()
    
    
    read_file_from_dags_bucket()
    
  2. Upload the read_file_from_dags_bucket.py DAG file to the bucket you created earlier. This will automatically create a graph with the same name in the Apache Airflow™ web interface.

  3. Open the Apache Airflow™ web interface.

  4. Make sure a new graph named read_file_from_dags_bucket has appeared in the DAGs section.

    It may take a few minutes to upload a DAG file from the bucket.

  5. To run the graph, click image in the line with its name.

Check the resultCheck the result

To check the result in the Apache Airflow™ web interface:

  1. In the DAGs section, open the read_file_from_dags_bucket graph.
  2. Go to the Graph section.
  3. Select the read_file job.
  4. Go to Logs.
  5. Make sure the logs contain the file content: {content} string, where content is the file content. This means the query was successful.

Was the article helpful?

Previous
Connecting to a Yandex Object Storage bucket with a bucket policy
Next
Working with Apache Airflow™ interfaces
Yandex project
© 2025 Yandex.Cloud LLC