이 제품은 선택한 Datadog 사이트에서 지원되지 않습니다. ().
이 페이지는 아직 한국어로 제공되지 않습니다. 번역 작업 중입니다.
현재 번역 프로젝트에 대한 질문이나 피드백이 있으신 경우 언제든지 연락주시기 바랍니다.
Join the Preview

Patterns is in Preview.

Request Access

Overview

Patterns automatically clusters your LLM application’s production traffic into meaningful topics, helping you understand what users are asking, identify coverage gaps, and diagnose failure modes.

How it works

Patterns uses text embeddings to group your application’s inputs into hierarchical topics. Topic labels are automatically generated using an LLM, giving you an interpretable view of production behavior without manual tagging.

When you run a pipeline, Patterns:

  1. Pulls LLM interactions from your production traffic based on your filter and sampling configuration
  2. Embeds interactions semantically and clusters them
  3. Names each cluster with an AI-generated label and summary using Datadog’s in-house models
  4. Organizes clusters into a parent-child topic hierarchy

Each topic shows its interaction volume, share of total traffic, and a coherence score — a measure of how semantically similar the interactions within the topic are to each other (0.0–1.0). Interactions that don’t fit any cluster are collected into an Outliers group.

Explore your Patterns

Read the summary metrics

The top of the Patterns page shows three metrics from your most recent run:

  • Total interactions: How many interactions were analyzed
  • Identified topics: The total number of distinct topics found, including parent and child topics
  • Classified: The percentage of analyzed interactions assigned to a named topic — interactions in Outliers count as unclassified

A high Classified percentage (above 80%) means the pipeline found meaningful structure in your traffic. A low percentage suggests high variance across interaction types or a filter that spans very different use cases.

The Patterns page displays traces grouped by topic.

The topic table provides a hierarchical view of all discovered topics. Each topic shows:

  • Topic name — auto-generated based on the interactions in the cluster
  • Summary — a plain-language description of what the topic represents
  • Interactions — count and percentage of total traffic
  • Coherence — a measure of how semantically similar the interactions within the topic are to each other (0.0 – 1.0)

Expand parent topics to see their sub-topics and examine specific areas of your application’s traffic.

Open a topic’s detail view

Click any topic name to open the detail view. Here, you can:

  • Read the summary of what this topic represents
  • View the scatter plot. Each dot is an interaction, plotted by semantic similarity. Tighter clusters mean higher coherence.
  • Browse the interactions table: real user inputs and outputs from production, with the sub-topic label and a confidence score for each
  • Navigate to child topics listed below the scatter plot
Topic detail view showing scatter plot of interaction embeddings alongside a table of interactions with topic labels and confidence score

Trigger a new run

You can trigger a new clustering pipeline run to re-analyze your production traffic.

  1. Click Run Pipeline.
  2. Configure your analysis:
    • Filter: Scope to a specific application, environment, or span type.
    • Sampling rate: Set what percentage of matching interactions to include. The pipeline processes up to 10,000 records per run; if your filter matches more than that, records are randomly sampled down to the cap.
    • Minimum Cluster Size (Advanced): Set the minium threshold for topic formation
  3. Click Run. The pipeline runs in the background and takes 5 to 10 minutes. You can close the page, or visualize the progress of the pipeline hovering on the status pill.

When the pipeline completes, the Patterns page updates with the run date, lookback window, and status.

Use topics to improve your application

Understand your production traffic

Use the topic list to see what are users actually doing with their application.

Use traffic percentage to identify your most common use cases. The parent-child hierarchy helps you move from a high-level pattern down to the specific sub-patterns underneath.

Find evaluation coverage gaps

Compare your topic distribution against what your golden datasets actually cover. Look at topics that represent high production volume but have no corresponding evaluation cases: this is where your test coverage has gaps, and where model regressions are least likely to be caught before they reach users.

Diagnose failure patterns

Scope your pipeline filter to spans with poor quality scores or failed evaluations, then run the pipeline. The resulting topic taxonomy shows which types of requests are failing most, giving you a structured way to prioritize fixes instead of debugging trace by trace.

Track how traffic evolves

Re-run the pipeline periodically and compare topic distributions over time. When a new topic appears near the top that wasn’t there last month, this indicates that your users have found a new use case (or a new failure mode).