The Measurement Blind Spots
- Shy B.T.
- Jan 18
- 3 min read
Most mobile teams don’t fail because they lack data, they fail because they trust the wrong parts of it or don’t really know what they’re looking at
After working with advertisers, DSPs, and internal growth teams, I’ve noticed a pattern - measurement blind spots rarely look like errors in your data, most of the time it's just a “good enough” dashboard or an “industry standard” setup coupled with decisions on attribution logic no one remembers making.

There are probably 20 blinds spots I could list, but here are the first 4 the came to mind while writing this post -
Blind Spot #1: Treating Attribution as “Truth” Instead of a Model
Attribution is not reality..... It’s a model of reality, with assumptions you baked into it when you setup your measurement.
When teams forget that, shit hit's the fan -
Dashboards turn into meaningless scoreboards
Discrepancies become political between teams and partner integrations
“Which number is correct?” replaces “What decision should we make?”
Different partners, windows, and engagement types can all be “right” and still tell different stories.
The danger isn’t disagreement, It's not knowing which story your organization is currently acting on.
Blind Spot #2: Assuming “Standard MMP Setup” = “Correct Setup”
Most apps running today will go with a default MMP configuration:
Default attribution windows & engagement priorities (7 days for Click, 24H for view)
Simple and non-strategic event mappings (SDK originating, simple purchase or progress events)
The most basic MMP SDK implementation their dev team agreed to do
A lack of comprehension regarding how the attribution logic functions
None of this is wrong by definition. , But none of it is neutral.
MMPs don’t know your internal business logic, and they rarely surface the full range of customization options unless you actively look for them or if you have a really good CSM.
Every attribution rule is a business decision in disguise and If you didn’t explicitly choose it, you’re still living with its consequences.
What could possibly go wrong? Here are a few examples:
UA optimizes for signals that don’t align with revenue timing
Retargeting claims credit the business doesn’t believe in
Product teams question marketing data but can’t point to why
You are actually sending wrong signals back to your ad networks, downgrading your performance
The real issue: No shared understanding of why the setup looks the way it does.
Blind Spot #3: No Measurement Ownership
One of the most common blind spots I see:
Each team optimizes their slice of the funnel perfectly.
UA optimizes CPI/M or early ROAS
Retargeting optimizes short-term lifts or dormant users
BI optimizes report consistency, data health and execution internal logic
Product works on alignment of the business logic with internal tools & dashboards
Each team will utilize different attribution links, access the data asynchronously from other teams, or simply activate or deactivate something without any notification.
Individually, each decision makes sense, collectively, they often cancel each other out or create a fucking mess in the data.
Measurement becomes fragmented, not because people are careless, but because no one owns the full measurement logic end-to-end.
Blind Spot #4: Confusing Ongoing Activity With Understanding
Dashboards update, numbers move and weekly reports are sent, and yet, when something breaks, no one knows:
Which assumption failed
Which rule is responsible
Who should own the investigation
A healthy measurement setup isn’t just observable - It’s explainable.
If your team can’t articulate:
Why a campaign is credited
Why an event is prioritized
Why a discrepancy is acceptable
Then you don’t have clarity - you only have data on dashboards running pointlessly.
A final thought
Most teams don’t need more tools, more dashboards, or more data.
They need a shared understanding of:
Which assumptions their measurement is built on
Where those assumptions help
Where they quietly limit better decisions
Looking under the hood doesn’t mean something is broken, It usually means the team has matured enough to question what it’s been taking for granted.
If this post raised more “we should probably talk about that” moments than clear answers - that’s not a problem, that’s usually the starting point.

Comments