Case Study: Diagnosing and Addressing a Metric Drop

An end-to-end framework for investigating MAU drops, targeting at-risk users, and making data-driven 'ship' decisions.

Remember the framework:

1. Set the Context for the Experiment


2. Design the Experiment

Monthly active users (MAU) is a key performance indicator (KPI) that measures the number of unique users who engage with a site or app within a month. It serves as a main metric for assessing a company’s overall health and user engagement, often influencing investor sentiment. While MAU is usually tracked using unique identifiers like emails or usernames, measurement methods vary across platforms, making direct comparisons between companies difficult.

Consider a ride-hailing service. Let us define Monthly Active Users as the number of users completing $\ge 1$ ride. Suppose the company observes a 7% drop in MAU.

1.

Is the drop natural?

You would slice the data to isolate the anomaly:

Is the drop resulted from external competitors?

If the drop is external, you will usually see it localized by geography or user behavior rather than a technical failure:

Formulate the hypothesis

Treatment

Hypothesis

Metrics

Phase 2: Experiment Design

Population

Deciding who to give coupons to (Eligible Users).

We shouldn’t give coupons to everyone. Giving them to guaranteed riders cannibalizes revenue, while giving them to completely lost users is wasted effort. I would frame our targeting as an optimization problem:

Unit of Randomization

  1. Geo-Experimentation (Cluster Randomization) + Synthetic Control This is the most robust way to test marketplace pricing and promotions.
  1. The Pragmatic Alternative: “Micro-Traffic” User-Level Randomization Geo-experiments are expensive and slow. In reality, tech companies often still use User-Level randomization for coupons, but with a strict mathematical constraint to prevent network interference.

Part 2: Design the experiment

1. Unit of randomization.

2. Statistical tests.

3. Power analysis and sample size calculation.


Part 3: The “Gotcha” Decision

Let’s say the result is statistically significant but not practically significant, would you ship it or not, and why?

I would not ship it.

Measuring Cost and Value (Trade-offs)

To measure the financial impact, the design needs to account for incremental lift.

Analysis

We are looking for the Incremental ROI. We would likely use a standard two-sample t-test on the mean revenue per user, or a z-test of proportions for the retention metric.

Implementation Note

Following the framework, I would launch this to a 1% traffic allocation initially. A bug in a pricing promotion can be incredibly costly, so we must verify the redemption logic and data pipelines for 48 hours before ramping up.


Part 3: The “Gotcha” Decision

Question: Let’s say the result is statistically significant but not practically significant, would you ship it or not, and why?

I would not ship it.