Welcome to Day 3 of Confident AI’s Launch Week.
Day 1 was Automated Error Analysis. Day 2 was Scheduled Evals. Today is the missing piece that makes both of those workflows actually sustainable.
Launch Week Day 3 (3/5): Auto-ingest traces into datasets and annotation queues.
The Most Valuable Data You’re Not Using
Every LLM team says they want “more real data”.
Then they ship to production… and their traces just sit there.
Not because they don’t care — but because turning traces into something useful is a surprisingly annoying workflow:
- Export traces from your observability system (or your internal logs).
- Normalize the schema so you can use them as a dataset (inputs, outputs, context, tool calls, metadata).
- Sample intelligently (because you can’t label everything).
- Route the right examples to humans for review (the ones that actually matter).
- Do it again next week because the world changed and your model drifted.
If this sounds familiar, it’s because most teams end up with one of two outcomes:
- They never build the pipeline, so they keep evaluating on stale, hand-curated datasets.
- They do build it, and it becomes a brittle ETL job that breaks the moment the product changes.
The Problem Isn’t Tracing — It’s The “Datasetization” Step
Tracing vendors are great at helping you inspect individual requests.
But your team doesn’t improve quality by inspecting individual requests. You improve quality by turning real production behavior into:
- Datasets you can run evals on (reproducible, versioned, comparable).
- Annotation queues your team can label (so you can learn what “good” and “bad” actually look like at scale).
So the question becomes, why is this step still manual in 2026?
Auto-Ingest on Confident AI
Auto-ingest lets you take production traces and automatically route them into:
- a dataset (for evals), and/or
- an annotation queue (for human review and error analysis),
…continuously.
You set it up once. It runs forever. No scripts. No cron jobs. No “export traces to CSV” rituals.
Here’s the workflow:
- Pick a trace source. Select the environment/project you want to ingest from.
- Define filters and sampling. Filter by route, tag, model, latency, cost, or any metadata you already attach. Add sampling so you only ingest what you can actually review.
- Choose destinations. Send the traces into a dataset, an annotation queue, or both.
- Map fields (optional). If you want a clean dataset schema, map trace fields into dataset columns.
- Done.
A Concrete Example
Let’s say you run a customer support agent. You want to:
- continuously collect real customer questions,
- label the failures (hallucinations, policy refusals, wrong actions),
- and re-run evals weekly on what users actually asked last week.
Your auto-ingest rule might look like this:
source:
environment: production
filters:
route: "/support/chat"
tags:
- "agent"
sampling:
strategy: "stratified"
by: ["model", "country"]
max_per_day: 200
destinations:
dataset: "support-prod-traces"
annotation_queue: "support-error-analysis"
schema_mapping:
input: "trace.input"
actual_output: "trace.output"
context: "trace.retrieved_context"
metadata:
trace_id: "trace.id"
model: "trace.model"
latency_ms: "trace.latency_ms"Now you get a dataset that stays fresh without you thinking about it, and an annotation queue that stays relevant without you manually curating it.
Why This Matters
This feature is less “shiny” than Day 1 or Day 2 — but it’s the one that makes them stick.
- Error analysis needs fresh failures. If your queue is built from last month’s exports, you’re learning the wrong lessons.
- Scheduled evals need living datasets. If your dataset is stale, your cadence is just a recurring illusion of control.
- Your best eval data is your real traffic. Auto-ingest turns production into a continuous feedback loop.
And the best part is that once traces are flowing into datasets and annotation queues, everything else becomes straightforward:
- you can label consistently,
- you can version datasets,
- you can track drift over time,
- and you can move from “we think it got worse” to “we know exactly what got worse, when, and why.”
What’s Next
This is Day 3 of 5. Two more launches coming — and as usual, it’s all about taking the workflows teams already do manually and turning them into infrastructure.
If you want to start auto-ingesting your traces today, sign up for Confident AI and set up your first ingest rule in a few minutes.
Do you want to brainstorm how to evaluate your LLM (application)? Ask us anything in our discord. I might give you an "aha!" moment, who knows?
The Eval Platform for AI Quality & Observability
Confident AI is the leading platform to evaluate AI apps on the cloud, with metrics open-sourced through DeepEval.

