RESEARCH / DATA TRACKING

Your Analytics Are Lying to You - Here's What to Fix

Dawid Jozwiak · · 13 min read

Why should you audit your analytics before doing anything else?

Because every decision you make downstream is only as good as the data it’s built on. In the Growth Recon framework, the Research stage exists to separate what you know from what you assume - and nowhere is that gap wider than in your analytics setup. Most companies treat their dashboards as truth. They aren’t truth. They’re a reflection of whatever some engineer configured eighteen months ago, filtered through attribution logic nobody fully understands, displayed in a tool that’s been miscounting events since the last website redesign.

Bad data doesn’t announce itself. It looks exactly like good data. The numbers go up and to the right, the reports are clean, and everyone nods along in the weekly meeting. Meanwhile, you’re optimizing for a metric that doesn’t reflect reality, spending budget on a channel that isn’t actually converting, and ignoring the one signal that would have told you the funnel was leaking at stage two.

The audit: what to look at first

Before you fix anything, you need to understand what’s broken. The analytics audit has four layers, and you work them in order. Skipping ahead means you’ll fix symptoms instead of causes.

Layer 1: Event accuracy

Every analytics tool works on events - page views, clicks, form submissions, purchases. The question is whether those events are firing correctly, consistently, and completely.

Start with your conversion events. Pick the three most important actions a user can take on your site - the ones that show up in your reports as KPIs. Now open your browser’s developer tools, perform each action, and watch the network tab. Did the event fire? Did it fire once, or did it fire three times because someone copy-pasted the tracking snippet into multiple templates? Does the event carry the right properties - the page URL, the UTM parameters, the user ID?

You’d be surprised how often the answer is “no.” A site redesign six months ago broke the form submission event on mobile. The thank-you page redirect strips UTM parameters before the conversion pixel loads. The SPA navigation doesn’t trigger pageview events on route changes. These aren’t edge cases. They’re the norm.

The fix: Build a tracking validation document. For every critical event, record what should fire, what data it should carry, and how to verify it. Then verify it manually on every major browser and device type. Automate this check if you can - there are tools that simulate user journeys and validate event payloads - but manual verification comes first, because automated checks only catch the failures you anticipated.

Layer 2: Data plumbing

Events fire correctly on your site. Good. Now follow the data downstream. Does it arrive in your analytics platform with the right properties intact? Does it flow correctly into your CRM? Does the integration between your ad platform and your analytics tool match conversion counts?

This is where most tracking setups quietly fall apart. The data leaves your site correctly but gets mangled in transit. A common example: your Google Analytics shows 200 conversions this month. Your CRM shows 180. Your ad platform claims credit for 260. Three systems, three different numbers, zero clarity on which one is right.

The discrepancies usually come from one of four places:

Timing windows. Your analytics tool counts a conversion at the moment it happens. Your CRM counts it when a sales rep changes a status field, which might be three days later. Your ad platform counts it within a 7-day or 30-day click attribution window. These aren’t errors - they’re different definitions of the same word.

Deduplication logic. One person submits a form, refreshes the page, and submits again. Your site fires two events. Does your analytics tool deduplicate? Does your CRM create two records? If your systems handle deduplication differently, your numbers will never match.

Filtering and bot traffic. Analytics tools filter out known bots. Your server logs don’t. Your form handler might count every submission, including the ones from spam bots in Eastern Europe. If you’re comparing raw form submissions against filtered analytics data, the gap is built in.

Cross-domain tracking. If your marketing site lives on one domain and your app or checkout lives on another, you need cross-domain tracking configured correctly. Most companies set it up once, confirm it works, and never check again. Then they add a subdomain for their blog, or move their pricing page, and the tracking chain breaks silently.

The fix: Pick one week of data. Reconcile every conversion across every system manually. Find every discrepancy. Document the cause. Then decide which system is your source of truth for each metric - and stop reporting the same metric from multiple systems with different numbers.

Layer 3: Attribution integrity

Attribution is where analytics goes from “slightly off” to “structurally misleading.” And most teams never realize it because the reports still look professional.

Here’s the scenario: a prospect reads your blog post via organic search. Two weeks later, they click a retargeting ad. A week after that, a colleague sends them a link to your pricing page. They bookmark it. Three days later, they type your URL directly and sign up. Under last-touch attribution, that’s a “direct” conversion. Under first-touch, it’s “organic search.” Under your ad platform’s self-reported attribution, it’s a paid conversion. All three are simultaneously true and deeply incomplete.

The Research stage doesn’t ask you to solve attribution - that’s a problem the industry has been arguing about for a decade. It asks you to document what your current attribution model can and can’t tell you, so that everyone making decisions based on that data understands the confidence level attached to it.

Run this exercise: Pull your top five acquisition channels by reported conversions. For each one, answer honestly:

  1. How much of this channel’s credit depends on the attribution model we chose?
  2. If we switched to a different attribution model, would this channel’s contribution change by more than 20%?
  3. Are we making budget decisions based on this channel’s reported performance?

If the answer to all three is “yes,” you have an attribution-dependent budget allocation - meaning your spend distribution is an artifact of your measurement model, not a reflection of reality. That’s not a reporting problem. That’s a strategic problem.

The uncomfortable truth about vanity metrics in attribution: Many teams avoid questioning attribution because the current model tells a story that justifies existing spend. The paid team likes last-touch because it makes paid look essential. The content team likes first-touch because it makes content look like the growth engine. Neither is lying. Both are selecting the lens that flatters their work. Your job in the Research stage is to strip that away and ask: what do we actually know, what are we inferring, and what are we guessing?

Layer 4: Coverage gaps

The first three layers are about fixing what’s there. This layer is about finding what’s missing.

Pull up your funnel. Map every stage from first visit to closed deal (or purchase, or activation - whatever your business’s definition of “converted” is). Now ask: at which stage transitions do we lose visibility?

Common blind spots:

The awareness-to-interest gap. You know how many people visit your site. You know how many fill out a form. But what happens between those two events? What pages do they visit? How many sessions before conversion? What content do they engage with? If your analytics is set up to track pageviews and conversions but nothing in between, you’re flying blind on everything that influences the decision.

The MQL-to-SQL gap. Marketing generates leads. Sales works them. But what happens in the handoff? How long do leads sit before first contact? What’s the response rate by lead source? If you don’t track the operational mechanics of the handoff, you’ll never know whether your conversion problem is a lead quality problem or a speed-to-lead problem. These have completely different solutions.

Post-purchase behavior. Most tracking setups end at the conversion event. The customer signed up - analytics job done. But churn is a growth lever. If you’re not tracking activation milestones, feature adoption, support ticket frequency, and engagement patterns in the first 30/60/90 days, you’re missing the signals that predict whether this customer will stick or leave.

Offline touchpoints. Events, phone calls, conferences, referral conversations. These don’t show up in your analytics because they don’t happen on your website. But they influence buying decisions - sometimes decisively. If your sales team closes 40% of deals that started with a conference introduction, and your analytics can’t see conferences, your data is systematically undervaluing your highest-performing channel.

The fix: Build a coverage map. Draw your funnel. At each stage, list what you can measure, what you can’t measure, and what you’re approximating. Share it with your team. The map itself is the deliverable - it tells everyone where the data is trustworthy and where it’s not.

Building a measurement framework that actually drives decisions

Auditing your analytics tells you where you stand. The measurement framework tells you what to do about it. A good framework has three tiers, and each tier serves a different decision-making need.

Tier 1: Operating metrics

These are the 4-6 numbers your team reviews every week. They should be leading indicators - metrics that predict future outcomes, not confirm past ones. If all your operating metrics are lagging (LTV, revenue, CAC), you’re making decisions with stale information.

Good operating metrics share three traits: they move fast enough to be actionable, they’re influenced by work your team is doing right now, and a change in direction triggers a specific response. “Website traffic” fails this test - it’s too broad to act on. “Conversion rate on the pricing page among visitors who read at least two blog posts” passes - because if it drops, you know exactly where to look.

Tier 2: Health metrics

These are the monthly or quarterly scorecards. LTV:CAC ratio. Churn by cohort. Return on ad spend by channel. Blended customer acquisition cost against target. These confirm whether your operating metrics are actually translating into business outcomes.

Health metrics are retrospective by design. Their job isn’t to tell you what to do this week - it’s to validate that what you did last quarter worked. When operating metrics look good but health metrics don’t improve, you have an efficiency problem: you’re optimizing the wrong thing. When operating metrics dip but health metrics hold, you might have a measurement problem, not a performance problem.

Tier 3: Diagnostic metrics

These live in reserve. You don’t report on them regularly - you pull them when something breaks. When your pricing page conversion rate drops, diagnostic metrics tell you whether it’s a traffic quality issue (new channel mix?), a UX issue (page load time on mobile?), or a positioning issue (competitor launched a lower price point?).

Diagnostic metrics require forethought. You need to have the tracking in place before the problem occurs, or you’ll be scrambling to instrument something while the fire is burning. During the Research stage, identify the diagnostic metrics you’d need to investigate the most likely failure modes for each operating metric. Set up the tracking now. Hope you never need it.

Assigning ownership

Every metric in your framework needs one owner. Not a team - a person. That person doesn’t have to be the one who moves the number. They’re the one who watches it, spots anomalies, and raises the flag. Without ownership, metrics become communal property, which means they become nobody’s responsibility.

The operating rhythm ties it together: weekly reviews of Tier 1 metrics, monthly reviews of Tier 2 metrics, and on-demand investigation using Tier 3 metrics when something looks wrong. This cadence prevents two failure modes - reviewing too often (reacting to noise) and reviewing too rarely (missing signals).

Common tracking failures and how to prevent them

After auditing dozens of analytics setups, the same failures appear repeatedly. Here are the ones that cause the most damage:

UTM anarchy. No naming convention for UTM parameters. One person tags a campaign as “spring_promo,” another as “Spring-Promo-2026,” a third as “spring promo.” Your analytics tool treats these as three separate campaigns. Multiply this across every campaign, every channel, and every team member, and your campaign-level reporting is fiction. Fix this with a UTM taxonomy document and a URL builder that enforces it.

Tag manager bloat. Your tag management container has 47 tags, 12 of which are broken, 8 of which are duplicates, and 4 of which are from vendors you stopped using two years ago. Every extra tag adds page load time and increases the chance of conflicts. Audit your container quarterly. If a tag doesn’t serve a current reporting need, remove it.

Sampling without disclosure. Google Analytics samples data when query volumes exceed certain thresholds. Your reports don’t mention this. So you’re making decisions based on data that represents 10% of your actual traffic, extrapolated by an algorithm. If you’re using sampled data, say so in the report. Better yet, configure your analytics to minimize sampling for your most important reports.

Cross-device blindness. A user researches your product on their phone during lunch, then converts on their laptop at the office. Without cross-device tracking (which requires authenticated sessions), these look like two different users - one who bounced and one who converted from direct traffic. Your mobile metrics look worse than they are. Your direct traffic metrics look better than they are.

Consent-driven data loss. Privacy regulations and browser-level tracking prevention mean you’re losing visibility on a growing percentage of your traffic. Cookie consent banners, ITP, ad blockers - each one removes a slice of your data. If you’re comparing this year’s numbers to last year’s without accounting for the growing measurement gap, your trend lines are misleading. Model the gap. Acknowledge it in your reports.

Running an A/B test you can actually trust

Your measurement framework is only as good as your ability to run clean experiments. Most A/B tests in marketing are statistically invalid - not because the math is hard, but because the execution is sloppy.

Three rules for tests that produce real signals:

Define the decision before you start. “If variant B beats variant A by at least 15% on conversion rate with 95% confidence, we’ll ship variant B.” Write this down before the test launches. If you decide what “winning” means after you see the results, you’re not testing - you’re storytelling.

Run to completion. Don’t peek at results on day three and call the test. Statistical significance requires a pre-determined sample size. Calculate it before launch. If your traffic volume means the test needs four weeks to reach significance, run it for four weeks. Early stopping inflates false positive rates dramatically.

Test one variable. If you change the headline, the button color, and the form layout simultaneously, you’ll know that “something” changed the conversion rate, but you won’t know what. Multivariate testing exists, but it requires much larger sample sizes. For most teams, sequential single-variable tests produce cleaner insights faster.

Where this fits in RECON

Data and tracking is one of four sub-areas within the Research stage of the Growth Recon framework, alongside ICP Mapping, the Language Audit, and the Adversarial Assessment. It feeds directly into The Source Doc - the single reference artifact that every downstream decision traces back to.

When you complete the data and tracking audit, you should have four deliverables: a validation document confirming your critical events fire correctly, a reconciliation showing where your systems disagree and why, an attribution confidence assessment that tells decision-makers what they can and can’t trust, and a coverage map showing where your funnel has blind spots.

These deliverables don’t just serve the Research stage. They’re load-bearing for everything that follows. In the Expose stage, you’ll choose channels and allocate budget - decisions that depend entirely on whether you can measure what those channels produce. In the Convert stage, you’ll optimize funnel steps based on data that’s only useful if the tracking is accurate. In the Optimize stage, you’ll run experiments that need clean measurement to produce valid results. And in the Navigate stage, you’ll build the operating rhythm that keeps the system running - which requires metrics you can trust week over week.

Fix your tracking now, during Research, and every stage after it gets more efficient. Skip it, and you’ll spend the next six months making confident decisions based on numbers that were never right. The cost of bad data isn’t bad reports. It’s bad strategy - and by the time you notice, the budget is already spent.