Merry Christmas! πŸŽ… πŸŽ„ 'Tis the season of OTT binges/marathons.

TL;DR: When TV shows are normalized by progress instead of episode number, viewer drop-off follows a similar early-dip / mid-plateau / late-rise pattern across platforms β€” with meaningful uncertainty.

The chart shows viewer drop-off across a TV series, measured by where you are in the show rather than by episode number.

Each series is normalized from:

  • 0% β†’ first episode
  • 100% β†’ final episode ('final' here refers to last episode available for a given show in the dataset)

Episodes are grouped into 20 progress bins, and the average drop-off probability is computed within each bin. Lines represent the four most common streaming platforms in this dataset (Netflix, Hulu, Prime Video, Disney+). Shaded regions show ~95% confidence intervals (standard error-based).

Why normalize?

Because episode 5 means very different things in a 6-episode miniseries versus a 30-episode procedural. Normalization lets us compare patterns of viewer behavior, not catalog length.

What stands out:

  • Early-series churn (β€œpilot cliff”) appears across platforms.
  • Mid-series stability varies.
  • Drop-off often rises again near finales, suggesting selective completion rather than universal binge-through.

Important note:

This chart is not being a grinch – saying Platform X is β€œbetter” or β€œworse.” It reflects episode-level behavior in this dataset only. Episodes within the same show are correlated, and the confidence bands indicate estimate stability β€” not causal differences or platform quality judgments.

Data πŸ“Š: https://www.kaggle.com/datasets/eklavya16/ott-viewer-drop-off-and-retention-risk-dataset

Made using βš’οΈ: pandas + numpy + Matplotlib

Posted by VegetableSense

Share.

Comments are closed.