Yandex Cloud
Search
Contact UsTry it for free
  • Customer Stories
  • Documentation
  • Blog
  • All Services
  • System Status
  • Marketplace
    • Featured
    • Infrastructure & Network
    • Data Platform
    • AI for business
    • Security
    • DevOps tools
    • Serverless
    • Monitoring & Resources
  • All Solutions
    • By industry
    • By use case
    • Economics and Pricing
    • Security
    • Technical Support
    • Start testing with double trial credits
    • Cloud credits to scale your IT product
    • Gateway to Russia
    • Cloud for Startups
    • Center for Technologies and Society
    • Yandex Cloud Partner program
    • Price calculator
    • Pricing plans
  • Customer Stories
  • Documentation
  • Blog
© 2026 Direct Cursus Technology L.L.C.
Monium
  • Getting started
  • Overview
    • Getting started
    • Basic terms
      • Getting started
      • Auto-instrumentation of LLM applications
      • Manual instrumentation of LLM applications
      • Viewing and analyzing LLM traces
    • Limitations
  • Access management
  • Pricing policy
  • Terraform reference
  • Release notes

In this article:

  • What is LLM monitoring
  • How it works
  • How to activate LLM monitoring
  1. Traces
  2. LLM monitoring
  3. Getting started

Getting started with LLM traces

Written by
Yandex Cloud
Updated at March 24, 2026
  • What is LLM monitoring
  • How it works
  • How to activate LLM monitoring

What is LLM monitoringWhat is LLM monitoring

When an AI agent acts unexpectedly, e.g., returns a weird response, calls a wrong tool, or does not call anything at all, it is difficult to figure out the issue through standard logs and traces. The key information gets lost among the attributes. Thousands of characters worth of prompts, several model calls, chains of tools – everything is mixed up.

Monium Traces LLM monitoring is a special trace viewing mode adapted for AI agent diagnostics. The interface is optimized for large texts, including prompts, system instructions, and model responses: this is the kind of data you usually need to understand why your agent has decided a certain way.

LLM monitoring runs on top of standard OpenTelemetry traces, so no dedicated infrastructure is required. If you have not configured the sending of traces in Monium Traces yet, start from Connection setup.

image

How it worksHow it works

LLM monitoring is based on OpenTelemetry Semantic Conventions for GenAI. According to this standard, all the model conversation data, including prompts, responses, system instructions, and tool descriptions, is delivered in special span attributes. Thanks to these attributes, Monium Traces highlights LLM-specific data and structures it for easy analysis.

There are two types of spans essential for LLM monitoring:

Generation span. Wraps a call to the model. This span's attributes store the following: conversation history with prompts (gen_ai.input.messages), model response (gen_ai.output.messages), model name (gen_ai.request.model), number of input and output tokens. The interface uses exactly these attributes to build a readable conversation feed.

Tool call span. Child generation span which wraps a call to an external tool selected by the agent. Contains the tool description, call parameters, and the execution result.

Note

If you are new to tracing, refer to Basic terms for descriptions of its basic entities: traces, spans, and attributes.

For a description of the LLM trace viewing interface, see Viewing and analyzing LLM traces.

How to activate LLM monitoringHow to activate LLM monitoring

Activate using the standard trace sending mechanism in Monium Traces. Follow the instructions in Connection setup. As an additional requirement, send traces with GenAI attributes: without them, the interface will not be able to correctly recognize LLM spans and present conversations in a user-friendly format.

There are two ways to add GenAI attributes to traces:

Automatically, if using the OpenAI SDK, OpenAI Agents SDK, LangChain, or another supported framework. Hook up the OpenTelemetry auto-instrumentation library for your framework. It will automatically create spans and add GenAI attributes with almost no changes to the code. This is the quickest way to get started, especially if you have already configured the sending of traces.

To learn more about auto-instrumentation of LLM applications, see Auto-instrumentation of LLM applications.

Manually, if you need full control over the trace data, or your framework is not supported by auto-instrumentation. Add GenAI attributes to spans manually as per the OpenTelemetry Semantic Conventions standard. This requires minor changes to the code but provides maximum flexibility.

To learn more about manual instrumentation of LLM applications, see Manual instrumentation of LLM applications.

Was the article helpful?

Previous
Incomplete traces
Next
Auto-instrumentation of LLM applications
© 2026 Direct Cursus Technology L.L.C.