Yandex Cloud
Search
Contact UsGet started
  • Pricing
  • Customer Stories
  • Documentation
  • Blog
  • All Services
  • System Status
    • Featured
    • Infrastructure & Network
    • Data Platform
    • Containers
    • Developer tools
    • Serverless
    • Security
    • Monitoring & Resources
    • AI for business
    • Business tools
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Start testing with double trial credits
    • Cloud credits to scale your IT product
    • Gateway to Russia
    • Cloud for Startups
    • Center for Technologies and Society
    • Yandex Cloud Partner program
  • Pricing
  • Customer Stories
  • Documentation
  • Blog
© 2025 Direct Cursus Technology L.L.C.
Yandex AI Studio
    • About Yandex AI Studio
    • Yandex Workflows
    • Quotas and limits
    • Terms and definitions
  • Compatibility with OpenAI
  • Access management
  • Pricing policy
  • Audit Trails events
  • Public materials
  • Release notes
    • General questions about models
    • AI assistants

In this article:

  • General questions about AI assistants
  • Tokens and billing
  • Working with indexes
  • Working with functions and tools
  1. FAQ
  2. AI assistants

AI assistant

Written by
Yandex Cloud
Updated at November 6, 2025
  • General questions about AI assistants
  • Tokens and billing
  • Working with indexes
  • Working with functions and tools

General questions about AI assistantsGeneral questions about AI assistants

How do I write effective instructions for an AI assistant?How do I write effective instructions for an AI assistant?

When creating an AI assistant, provide a detailed plain-language description of what your assistant should do in the instruction field. For example:

You are a support specialist. Your task is to answer questions on the topic. 
Stick strictly to the context.
If the answer is not in the context, simply state that, without making assumptions.
Keep your responses short but informative.

Does the assistant wait for the model to complete its response before proceeding to the next event?Does the assistant wait for the model to complete its response before proceeding to the next event?

Yes. A single assistant run triggers a single model call. The model receives your prompt, invokes the required tools, waits for the results, and then generates a response.

Can I cancel a request and interrupt an ongoing model response?Can I cancel a request and interrupt an ongoing model response?

No, you cannot cancel a submitted request. To limit the model's response length, use the maxTokens parameter.

Can I send multiple requests to the same thread from different assistants concurrently?Can I send multiple requests to the same thread from different assistants concurrently?

Yes, there are no restrictions on concurrent use.

How do I prevent the transfer of context from the previous request?How do I prevent the transfer of context from the previous request?

You can specify the number of thread messages to send to the model with each new request using the thread_num_messages parameter. This way you can control how much context to use for the response, including messages from both the user and the model.

You can also create a new thread for each request and delete the old one.

Can the assistant detect images in text files?Can the assistant detect images in text files?

We plan to add support for image input.

Does AI Studio have native monitoring and quality assessment tools?Does AI Studio have native monitoring and quality assessment tools?

We are actively developing AI Studio and plan to introduce monitoring and quality assessment tools.

Which languages does the AI assistant support?Which languages does the AI assistant support?

The main YandexGPT Pro languages are Russian and English. Soon, we will release new large open-source models that will work well with other languages.

What are the differences between the RC and Latest versions?What are the differences between the RC and Latest versions?

We are continuously enhancing the response quality and features of our models. Once internal metrics show the new model version is ready, we publish the updates to the RC branch for testing and notify users in our release notes.

Learn more about the model lifecycle here.

Tokens and billingTokens and billing

How do I spend tokens when using the AI assistant?How do I spend tokens when using the AI assistant?

Just like with generative models, you pay for both request and model response tokens. This also includes context from the knowledge base and chat history. You can use the maxTokens parameter to limit the number of tokens in the model's response.

How can I estimate the number of tokens in text files beforehand?How can I estimate the number of tokens in text files beforehand?

To estimate the text size in tokens, use a tokenizer. Learn more in Estimating prompt size in tokens.

Do I get charged for the model's response to a request that triggered the ethics filter?Do I get charged for the model's response to a request that triggered the ethics filter?

No, you do not get charged for such responses.

Do I get charged for creating search indexes, uploading files through the Files API, or storing files and indexes?Do I get charged for creating search indexes, uploading files through the Files API, or storing files and indexes?

Note

We do not recommend using AI Assistant API in new projects. To create AI agents, use the Responses API.

Creating or storing files, threads, or indexes is free of charge.

Working with indexesWorking with indexes

How do I set up a knowledge base for my AI assistant?How do I set up a knowledge base for my AI assistant?

To set up a knowledge base for your AI assistant:

  • Create your documents as Markdown files.
  • Keep as much information in plain text as possible.
  • Remove any footnotes and comments to make the content concise and clear.
  • If your documents contain tables, convert them to Markdown format.
  • Make sure each text chunk is big enough to include the largest tables from your documents.

How can I build an assistant that can search documents, use them to generate responses, and provide source links?How can I build an assistant that can search documents, use them to generate responses, and provide source links?

Upload your documents to the assistant's search index. The assistant will then return answers that include both links to the source documents and exact text chunks the model used to generate the response.

Learn more in Creating a RAG assistant with the Vector Store tool.

How do I add, update, or delete documents from the index?How do I add, update, or delete documents from the index?

To add a new document to your index, use the append operation.

To update a document, you need to rebuild the entire index.

How do I upload files directly from an object storage without saving them to my computer?How do I upload files directly from an object storage without saving them to my computer?

You can create a script that pulls data directly from your storage and uploads it to your AI assistant using the sdk.files.upload_bytes() method.

What method should I use to upload files from RAM?What method should I use to upload files from RAM?

The sdk.files.upload_bytes() method works best for this kind of upload.

What happens if I delete a file used in the index?What happens if I delete a file used in the index?

After the search index is created, it does not rely on files anymore. If you delete a file, the assistant will keep working correctly, but its responses will not include the link to that file.

Can I set up my AI assistant to provide no links when the index has no relevant content?Can I set up my AI assistant to provide no links when the index has no relevant content?

Currently, this setting is not supported.

Can I connect multiple indexes to the same assistant?Can I connect multiple indexes to the same assistant?

Currently, you can only connect one search index to each assistant.

You can either update the assistant with a new index or create two assistants, each with its own index.

Can I select a specific file for my AI assistant to search for an answer?Can I select a specific file for my AI assistant to search for an answer?

If you want your AI assistant to answer based on a specific file, create a dedicated search index for that file and update the assistant to use it.

You can also create multiple assistants with different indexes.

Working with functions and toolsWorking with functions and tools

Can the model work with a toolchain?Can the model work with a toolchain?

To use a toolchain, you need a wrapper, such as LangGraph. For an example of how to work with a toolchain using LangGraph, see this repository.

How can I improve the quality of function calls?How can I improve the quality of function calls?

To improve the quality of function calls:

  • Add usage examples to the function descriptions.
  • Set clear conditions for function calls in the instructions for your AI assistant.
  • Expand parameter descriptions by specifying typical values.

How do I manage function selection and calling conditions in the assistant?How do I manage function selection and calling conditions in the assistant?

Implement fallback logic to call a function when it is not explicitly selected but the input contains relevant keywords. Use an external intent analyzer to improve recognition quality.

Was the article helpful?

Previous
General questions about models
Next
Error codes
© 2025 Direct Cursus Technology L.L.C.