The Metrics That Actually Matter for Product Managers
Most product metrics are vanity. Here are the ones I actually track — and why.
Most product metrics are vanity. Here are the ones I actually track — and why.

I've seen teams drown in dashboards. 50 metrics, beautiful charts, weekly reviews — and nobody can tell you if the product is actually doing well.
Metrics aren't the goal. Understanding your product is. Here's what I actually track.
Most metrics fall into two traps:
Vanity metrics: Numbers that go up but don't mean anything. Page views. Registered users. App downloads. They feel good but don't tell you if people are getting value.
Lagging metrics: Revenue. Churn rate. NPS. By the time these move, it's too late to react. You need leading indicators.
Depends on the product, but here's my core set:
I don't care how many people signed up. I care how many people used the product this week.
For a payments platform I worked on: 10K registrations meant nothing. 2K active merchants processing transactions meant everything.
Not just "what's our retention rate" but "how does retention change over time by cohort?"
Plot it:
Is that curve flattening or still dropping? If it flattens, you have a core user base. If it keeps dropping, you have a leaky bucket.
I check this monthly. Changes in the curve tell you more than any single number.
What percentage of new users reach the "aha moment"?
You need to define what activation means for your product:
If activation is low, fix that before anything else. Acquisition doesn't matter if people don't stick.
Which features are actually being used?
I've killed features with 2% adoption. I've doubled down on features with 60% adoption. Usage data should drive roadmap decisions.
Check:
Not a single NPS number. A continuous stream:
I review support ticket themes weekly. Spikes tell you something's wrong before metrics do.
Page views / session time — Unless you're a content business, these don't matter. Someone could spend 30 minutes frustrated trying to do something simple.
Story points / velocity — Team capacity metric, not product success metric. Delivering 100 story points of the wrong thing is worse than 20 points of the right thing.
Total revenue — Too lagging. By the time revenue drops, the damage happened months ago. Track leading indicators instead.
One page. Five numbers:
Plus one qualitative section: notable feedback from this week.
Takes 10 minutes to review. Tells me everything I need to know about whether the product is healthy.
Numbers don't lie, but they can mislead.
Example: Activation rate went up 15%. Celebration? Turns out we lowered the bar for what counted as "activated." Users weren't actually getting more value.
Example: DAU spiked after a press mention. Looked great. But retention for that cohort was terrible. Vanity traffic, not real growth.
Always ask: "What could make this number move without actual improvement?"
Behind all the numbers: Are users getting value?
If yes, other metrics will follow. If no, no amount of growth hacking will save you. I've seen products with great metrics die because the core value wasn't there. I've seen products with terrible metrics survive because users genuinely loved them.
Track the numbers. But don't mistake the map for the territory.