ToolStack
Quality Metric

Error Rate

Error rate measures the proportion of user interactions, API calls, or system events that result in an error. For product teams, it's a leading indicator of user experience quality — rising error rates precede churn and support volume spikes. For engineering teams, it's a primary SLI (Service Level Indicator) within SLOs. Tracking error rate by feature, endpoint, and user segment enables fast, targeted remediation.

Formula
Error Rate = (Number of error events) ÷ (Total events in same window) × 100

Note: Track separately: client-side errors (JS exceptions, crashes), server-side errors (5xx responses), and business logic errors (invalid state transitions). They have different causes and different impact on users.

Healthy range

API error rate < 0.1% (99.9% success rate); critical path error rate < 0.01%

Warning signs

Error rate > 1% on core user flows will measurably increase churn and support volume

Benchmarks by segment

SegmentBenchmark
Consumer apps (industry standard)Crash-free sessions > 99.5%
B2B SaaS APIs (tier 1)< 0.1% 5xx error rate on core endpoints
Payment flows< 0.01% error rate (errors here directly lose revenue)
Data pipelines< 0.5% processing error rate

How to improve Error Rate

1

Set up real-time error alerting with severity tiers — P0/P1 errors page on-call; P2/P3 go into a triage queue

2

Build error budgets: define an acceptable monthly error budget per endpoint; when the budget is exhausted, stop new feature work and fix reliability

3

Add structured error logging so every error is tagged with user segment, feature, and environment — makes root-cause analysis 10× faster

4

Run chaos engineering (controlled fault injection) to find reliability weaknesses before users do

Common measurement mistakes

!Tracking only 5xx errors and missing 4xx errors that indicate broken user flows or bad API contract design
!Treating all errors equally — a 500 on an avatar upload is less critical than a 500 on the checkout page
!Measuring error rate without correlating to user impact: some errors are silent (retry succeeds); others are fatal (user loses work)

Tools for measuring Error Rate

#1
Amplitude
4.5Free tier

Best-in-class behavioral analytics with powerful event segmentation, funnel analysis, and retention charts that go far deeper than Google Analytics

#2
Mixpanel
4.6Free tier

Best-in-class event-based analytics with intuitive funnel, retention, and flow reports that surface actionable insights quickly

#3
FullStory
4.5Free tier

Best-in-class autocapture technology — captures every click, scroll, and interaction without manual event tagging, enabling retroactive analysis on historical data

#4
PostHog
4.6Free tier

All-in-one product analytics platform combining analytics, session replay, feature flags, A/B testing, surveys, and a data warehouse — replacing multiple point solutions

#5
Heap
4.4Free tier

Autocapture eliminates the need for manual event instrumentation — every click, pageview, and form interaction is tracked automatically from day one

#6
Statsig
4.7Free tier

All-in-one platform combining feature flags, A/B testing, product analytics, session replay, and web analytics — eliminating the need for separate tools

Frequently Asked Questions

What is an error budget?

An error budget is the maximum acceptable error rate for a service over a rolling window (typically 30 days). If your SLO is 99.9% success rate, your error budget is 0.1% — that's 43.8 minutes of downtime or errors per month. When the budget is consumed, reliability work takes priority over new features.

How do I track error rate without overwhelming my team with noise?

Use severity tiering and deduplication. Group repeated identical errors into a single alert. Set minimum thresholds (e.g. only alert if > 10 errors in 5 minutes). Tools like Sentry, Datadog, and Rollbar handle this automatically.

Related metrics

p99 LatencyFeature VelocitySupport Ticket Volume