🎯 The Deceptive Nature of Averages | Why Packet Loss Hides in Time Series Graphs

Ronald BartelsRonald Bartels
3 min read

In network diagnostics, the devil is almost always in the detail — and nowhere is this more painfully true than when looking at time series graphs of packet loss. Engineers and support staff often rely on these graphs to evaluate network health, but if you're not careful, you’ll miss the moments where real users were screaming — all because of how averages behave.


📉 The Case of the Disappearing 2%

Let’s paint a picture. Imagine a link suffers 2% packet loss for five minutes around 10:15 AM. It's just enough to break video calls, stall remote desktop sessions, or make voice calls sound like a drowning robot 🫠.

But when you look at the daily graph of that link — neatly sampled every 5 or 15 minutes, perhaps even worse — the loss appears as a mere blip, if it's even noticeable at all. Why?

Because of averaging. Over a 24-hour period, five minutes of 2% loss becomes:

(5 minutes × 2% loss) / (1440 minutes in a day) ≈ 0.007% averaged loss

That’s right — it barely even registers as a bump on the graph. The packet loss that caused users to log support tickets, restart routers, and maybe even threaten to switch providers is statistically smoothed into irrelevance.


🔬 Why You Can't Rely on Coarse Views

The problem gets worse when using monitoring tools that:

  • Only sample at 5–15 minute intervals,

  • Show only daily or weekly rollups,

  • Don't allow zooming into minute-level granularity.

This makes short bursts of significant impact invisible in the wider statistical view.

🔍 Think of it like looking at a heart monitor with one reading per hour. You’ll never catch the heart attack.


🛠 Why This Matters for Diagnostics

When trying to explain a network issue to a customer or isolate root cause:

  • You need access to high-resolution telemetry, especially around the time the user experienced an issue.

  • Diagnostic tools must allow you to zoom in on the timeline — ideally down to 1-minute or even second-level intervals.

  • Dashboards that only display smoothed daily averages are effectively gaslighting your support team.

You’ll end up in the classic trap:

"The graphs look fine" 😐
While users are still complaining about problems that are very real.


💡 Best Practices

To avoid this:

  • Implement tools that store and visualise short-term, high-frequency metrics (e.g., every 30 seconds).

  • Use alerting systems that trigger on consecutive small bursts of packet loss (e.g., 2% loss over 2 minutes).

  • Design dashboards with drill-down capability so engineers can inspect incidents at the correct time scale.

  • Combine SLA-focused metrics (loss, jitter, latency) with application-aware insights (e.g., VoIP quality scores).


🧠 Wrapping Up

When it comes to network health, not all metrics are created equal, and not all packet loss is equal. Averages may hide more than they reveal — and if your tools can’t see the spike, you’ll miss the story.

So next time a user says, “The internet broke for 5 minutes,” don’t just look at the 24-hour graph and say it looks fine.

Zoom in. The truth is waiting.

10
Subscribe to my newsletter

Read articles from Ronald Bartels directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Ronald Bartels
Ronald Bartels

Driving SD-WAN Adoption in South Africa