Skip to main content

Data Lake

Query all your pipeline runs to analyze sample processing turnaround times, costs per pipeline execution, error correlations and performance trends across projects.

How It Works

We ingest all your pipeline telemetry, metrics, and logs into a multi-tenant data lake. You can then use the Tracer web interface to query this data to analyze performance, identify bottlenecks, and optimize your pipelines. Identify underperforming steps, track which projects need attention, and monitor how efficiency changes over time. Tracer allows you to export all pipeline telemetry, metrics, and logs to your data lake for custom analysis and long-term retention. You can calculate cost per sample, measure run durations, and compare throughput across pipelines to identify optimization opportunities. Tracer also makes it easier to spot regressions, monitor efficiency trends, and inspect error types so teams can locate bottlenecks, prioritize fixes, and improve their workflows using real execution data.

Key Analytics

Pipeline Success Rate Analysis

Track overall pipeline health with metrics such as success rate, total executions, and average turnaround times. Pipeline Success Rate Analysis

Most Expensive Pipelines

Identify your most expensive pipelines for any time window and surface the average cost per run to understand spend drivers. Most Expensive Pipelines

Pipeline Resource Utilization

Spot pipelines that consistently underuse allocated compute, helping you pinpoint wasted capacity and optimize resource allocation. Pipelines Underutilising CPU

Longest Tool Execution

Surface the tools or steps that contribute the most to overall runtime, including the exact commands, so you know where to focus optimization efforts. Longest Tool Execution