Attribution Truths from (un)Common Logic Analysts

Marketing attribution promises the moon and then hands you a telescope. Most teams learn this the long way, by pinning a quarter’s spend to a glittering dashboard that can’t survive a skeptical question. Having audited dozens of programs and built a few from the ground up, our analysts at (un)Common Logic have collected a set of hard truths that make attribution more useful, more trustworthy, and more likely to change real budgets.

The most important thing to remember: attribution is a decision support system, not a court of law. It should narrow uncertainty enough to move money with confidence, not claim perfect credit assignment.

image

The promises that cause rework

Attribution slides well in a pitch deck. A beautiful Sankey diagram seems to show the exact path to revenue, neatly labeled by channel and touch. That picture invites dangerous assumptions.

First, many tools smuggle policy choices in as if they were facts. A 7 day click window or 1 day view window is not a law of nature. It is a choice about how long influence lasts for your brand, your product, your audience. If someone else picked it for “industry standard” reasons, you just inherited their business model and conversion cycle.

Second, identity resolution is never complete. Cookies expire. Apps wall off data. Email matches skew toward loyal users. A last touch record stands in for a journey, especially for buyers who never clicked an ad. If your mix leans into video, CTV, or upper funnel social, click trails will undercount it by design. A tool that “solves” this with wide view windows often papers over the problem and quietly boosts every platform’s self-reported numbers.

Third, most dashboards are calibrated to be consistent with themselves, not with cash. They can align conversions between platforms and analytics, then leave you 18 percent off from the bank account. The best ops teams tie attribution back to dollars in the ledger, after refunds, chargebacks, and cancellations. The best analysts accept that some portion of spend remains unattributed in any one method and then measure the gap instead of pretending it is not there.

What changes, and what stubbornly does not

Attribution is more complicated than it was five years ago. Privacy changes on iOS, shorter cookie lifetimes, and platform reporting gaps raised the noise floor. Even so, the backbone of trustworthy attribution has not shifted.

What changes:

    Identity stability ebbs, so deterministic matching alone undercounts reality. You will need modeled conversions, conversion APIs, and direct platform integrations that do not rely only on browser storage. Channels fragment. Retail media, streaming, influencer, and affiliates each bring different data grains and lag patterns. You cannot shoehorn them into a one size clicks table. Platform self-attribution expands. Walled gardens got better at measuring inside their walls and worse at sharing outside. Their numbers rise with looser rules. You will need independent checks.

What does not:

    Causality still demands a counterfactual. Without a believable “what would have happened anyway,” you are reading tea leaves. Diminishing returns remain. The second dollar almost never performs like the first, no matter what a linear or last click line says. Decision cadence still matters more than theoretical accuracy. A decent answer available every two weeks beats a pristine answer that arrives after planning season.

Five truths we have seen hold up

    Measurement without a test harness drifts. Models get stale. Platform tags break. Creative shifts change who sees your ads. If you do not run planned holdouts, geo splits, or PSA swaps at regular intervals, your attribution will quietly self validate. Your data design choices change your ROI more than your model choice. Whether you dedupe conversions across platforms, standardize UTMs, define channel taxonomy, and set sensible conversion windows matters more than picking Shapley over Markov for multi touch. A sloppy foundation can swing channel ROI by 30 percent. Clean plumbing reduces the swing to within a tolerable error band. Use two lenses, not one. Combine a top down model like MMM, anchored in spend and outcomes, with bottom up journey data where you have identity. Each one contradicts the other in useful ways. When they disagree, you learn where the uncertainty hides. When they agree, you can move money faster. Time and geography beat most precision tricks. A well designed geo experiment with 12 to 30 test markets, balanced by baseline sales, often isolates incremental lift better than a click based model with twice the features. You learn how spend scales, not just how it sequences. Confidence intervals belong on budgets. If your MMM says paid social returns 2.3 to 3.1 ROAS at current levels, plan with the low end for safety and with the mid for growth. Writing ranges into plans makes downstream reporting honest. It also trains executives to expect movement, not a single heroic number.

Anatomy of a foundation you can trust

Attribution stands or falls on data contracts. Not just legal ones, but practical agreements across teams. The cleanest implementations we have seen looked unglamorous on day one and saved months of churn later.

Start with identity and events. Pick a primary key you control, even if it only shows up part of the time. For web, lean on first party cookies and server side tagging to preserve session logic. For app, stabilize on device IDs where allowed and your internal user ID otherwise. Never let a platform pixel fire a conversion event that your source of truth does not also log. When finance asks why Meta shows 12,430 purchases and your warehouse shows 11,200, you should have an exact reconciliation path, not a shrug.

UTM governance sounds dull. It is a lever. We once found five spellings of the same channel in a client’s links, which scattered revenue across rows that looked unrelated. A two page guide, a required parameter set, and a weekly audit script turned their messy reporting into a coherent picture. The resulting change in paid search optimization lifted non-brand ROAS by roughly 15 percent within a month, not because the channel changed, but because the feedback loop finally told a consistent story.

Define deduplication rules that reflect how users actually buy. If someone clicks a paid search ad, then a Facebook ad, then purchases after an email, what gets credit? Some teams default to “last touch wins” and call it done. More mature teams express a policy: paid channels split credit if they occur within seven days before the last owned touch, which claims only a fixed share unless it started the path. You can argue the details, but the point is to set rules on purpose, then implement them in both your reporting layer and your optimization tooling so the incentives align.

Set conversion windows with evidence. For a $900 product that buyers research over weeks, a 1 day click window pretends that money falls from the sky. For a $12 impulse buy, a 7 day view window double counts ambient behavior. Use cohort curves from your own data. If 85 percent of attributed paid search conversions arrive within three days of the click, that is your starting point, not a vendor’s default.

Finally, document channel taxonomy and touch rules. If influencer drives traffic through trackable links sometimes and brand mentions other times, split it into affiliate versus awareness subchannels. That separation lets you pay partners fairly and still defend your incrementality math.

Tests that settle arguments

The fastest way to put an attribution conversation on firm ground is to run a clean, interpretable experiment. This is less about statistical theater and more about making trade-offs visible.

Geo experiments bring power without needing user level joins. Pick matched markets with similar baselines, seasonality, and competitive pressure. Assign half to hold spend steady and half to increase by a defined amount. Run long enough to let media scale, often 4 to 8 weeks for retail and 8 to 12 for subscription. Keep leakage low by ensuring creatives differ or flights are confined to the geo. The output is a lift estimate and a scaling curve. We often target an 80 percent power to detect a 5 to 15 percent lift, which for mid sized advertisers means total spend in test geos large enough to move weekly revenue by a few points.

PSA swaps help answer view based questions. If you wonder whether your CTV partner’s view throughs are real, swap your ads for public service announcements in a random subset of spots while keeping the buy otherwise identical. Watch downstream site traffic and branded search in treated versus control footprints. A null result here tells you to reinvest elsewhere, even if platform reports glow.

Holdouts clarify CRM and retargeting. Withhold a well defined segment from email or retargeting for a fixed period. The revenue delta, adjusted for any spillover, tells you the incremental impact. Many teams are surprised to find that heavy retargeting of existing purchasers drives vanity metrics while doing little for net revenue. Savings from dialing back frequency often fund more prospecting.

Audience splits can adjudicate algorithmic preferences. If your multi touch model favors upper funnel social but paid search is the proven closer, split prospecting audiences and grow both for a month. Let revenue per marginal dollar decide. Then update your model priors with the observed lift so the tool learns along with you.

The key is not to test everything. Test the levers that would change next quarter’s budget. Publish protocols as one page memos that state sample sizes, windows, success criteria, and who signs off. When the test ends, show raw numbers and context, not just a verdict.

MMM that operators actually use

Media mix modeling lost some fans when fast moving teams met six month academic projects. That is a pity, because a lightweight MMM can pay for itself in budget flexibility alone.

Start with the basics that matter for decisions. Include weekly spend by channel, outcome events by market or region, price and promo flags, known seasonality, and exogenous factors like weather only if your category is sensitive to it. Model diminishing returns and adstock. If you cannot explain the meaning of each term to a budget owner in plain language, strip it until you can.

Bayesian approaches shine here because they handle uncertainty with grace. A prior that says paid social likely saturates faster than paid search is not a bias if you can justify it with tests or historical response curves. Use experiments as anchors. If a geo lift study shows a 12 percent incremental return for YouTube in the Midwest, set the prior or constrain the slope so the model does not deviate wildly without strong evidence.

image

Refresh monthly, not annually. Each refresh should ingest new spend and outcomes, reestimate curves, and output updated response at the margin. The most useful output is not channel credit, it is a budget optimizer with a credible band. If it says another 50,000 dollars in non-brand search yields 1.8 to 2.2 ROAS next month and another 50,000 in TikTok yields 1.4 to 1.9, you can move money with eyes open.

Beware of https://jaredihnw225.image-perth.org/turning-insights-into-tests-with-un-common-logic false precision. An MMM that explains 92 percent of variance on in sample data and then flops in the real world is a common failure. Overfit hides when you pack in too many correlated channels or let promo flags “explain” normal movement. Keep models lean and train them to forecast next month, not the last three years.

Clickstream models without delusion

Multi touch attribution still helps operators steer creative and journey tactics, as long as you do not pretend it captures everything.

Start with a sparse set of touches that you trust. Paid media clicks with solid UTMs, owned channel sends and clicks, site referrers you recognize, and partner traffic where contracts demand transparency. Do not cram in every impression event you can collect. There is a point where volume becomes anti signal.

Pick a method that fits your question. Shapley values support fair division when multiple touches collaborate. Markov chains help you see which paths collapse when you remove a channel. Logistic regression with time decay lets you control for user features or segments when you have them. No method rescues bad windows or sloppy dedupe. Get your policy right, then the math.

Use modeled conversions with care. Conversion APIs can restore lost signal from browsers, but they introduce a second source of truth. Reconcile modeled and observed events weekly. If modeled events creep above a set ratio, dig in. We have caught misfiring server tags that silently doubled view throughs in one partner but not others. A simple control chart saved a quarter’s reporting.

Finally, align optimizations to what the model believes. If your MTA devalues last click and rewards assist touches, make sure your bid strategies and creative objectives reinforce that. Teams often end up with split personalities, reporting one model while optimizing to another, and then wonder why results stall.

A practical example at operator speed

A mid market DTC retailer, average order value around 85 dollars, came to us with flat revenue despite rising spend. Their dashboard said paid social drove half of conversions by view through, while last click analytics crowned brand search. Finance saw gross profit stuck.

We tightened the foundation first. UTMs were stabilized. Email conversions were deduped against paid channels with a simple policy. Conversion windows were reset to 3 day click and 1 day view for paid social based on their own cohort curves. Modeled conversions were included, but capped to a sensible share of total if identity dropped.

Then came a geo test. We split 20 markets by historical sales volatility and competitor density. Ten markets increased paid social prospecting by 30 percent. Ten held steady. We ran six weeks, staying within creative norms to avoid novelty effects. Lift landed at 8 to 12 percent in test geos, with stronger response in markets that skewed younger. The platform reported more, as expected. The test told us what mattered.

An MMM refresh folded that lift into priors and suggested that at current levels, another 100,000 dollars in paid social would likely return 1.6 to 2.1 ROAS, while the same in brand search would return 1.2 to 1.5 due to saturation. Meanwhile the MTA, now on cleaner data, showed that non-brand search played a bigger assist role than their last click picture allowed.

We moved budget in two waves, first 10 percent, then another 10 after three weeks if guardrails held. Revenue responded within the test bounds. Gross profit ticked up. Not a miracle, just a series of grounded steps, all of which we could defend in a room with finance.

What to do this quarter

    Write and adopt a one page attribution policy. Cover windows, dedupe rules, channel taxonomy, and identity logic. Get marketing, analytics, and finance to sign. Schedule one decisive experiment. Pick the argument most likely to change budget, design a geo or holdout test, and set a date, power, and success yardstick. Stand up a monthly MMM refresh with tight scope. Spend by channel, weekly outcomes, adstock, diminishing returns, and a budget optimizer that outputs ranges. Clean the clickstream. Fix UTMs, remove untrustworthy touch types, and reconcile modeled versus observed conversions with a weekly control chart.

Edge cases that need judgment, not templates

Subscription businesses live in the land of lag. Trial to paid conversion can stretch across weeks. Retargeting may boost trial starts without moving net paid conversions. Your primary metric should tie back to downstream value, not front door vanity numbers. Cohort based MMM, where outcome is 28 or 56 day revenue from a signup cohort, beats simple signups in your objective function.

B2B cycles bring low volume and long paths. You will not get robust multi touch paths for deals that close in 4 to 12 months. Lean harder on controlled experiments at the account or region level, and on lift proxies like branded search volume or SDR connect rates. Bring qualitative feedback from sales into your priors, then verify with periodic tests.

Marketplaces have two sides. Ads that drive buyers and ads that recruit sellers interact. A campaign can appear to have weak direct ROAS and still be pivotal if it balances the ecosystem. Build a balancing constraint into your optimizer. Accept that some “spend” is maintenance, not acquisition, and measure health with ratios like buyers per active seller.

Mobile apps run into SKAdNetwork and privacy walls. ID matching across paid channels is sparse. Lean into geo lifts and on device experiments. Combine daily active users, retention curves, and modeled conversions with cautious priors. Do not backfill view throughs to hit goals. If you cannot measure a partner credibly, either isolate it with a test or cut it.

Retail media sits between sales activation and shopper marketing. Units sold on a retailer’s site may spike from a campaign that cannibalizes organic. The cleanest read comes from test versus control at the SKU or banner level within the retailer’s experiment framework. Pull that into your cross channel picture with care, since retailer definitions may disagree with yours.

What good looks like in numbers

Data freshness within 24 hours for clickstream events keeps operators confident and lets you sense breaks quickly. Identity match rates for deterministic joins vary, but 60 to 80 percent on known users for web and app is a practical target in many categories. Where you cannot match, model conservatively and disclose the share of modeled conversions weekly.

For MMM, an out of sample forecast error in the 5 to 15 percent range at the weekly level is realistic for mid sized businesses. If you are under 5 percent, you may be overfitting. If you are over 20 percent, simplify and retrain. Present channel ROAS as ranges, not points, and make spend recommendations with explicit marginal bands. Executives respect plans that acknowledge uncertainty more than they trust a single shiny number.

For experiments, aim for 80 percent power and a minimum detectable effect that ties to business goals. If a 5 percent lift pays back the test and changes budget, design for that. If you need 20 percent lift to care, set the test up to detect it quickly or skip it and move on.

For governance, track the share of conversions that reconcile to the source of truth. If more than 10 to 15 percent sit in a gray bucket of “platform only,” you need to review tagging, windows, or modeling caps. Weekly audits, even if light, catch drift before it derails a quarter.

Bringing it all together

The workable pattern we return to looks simple from the outside. A stable data foundation with explicit rules. One top down model that respects reality and returns ranges. One bottom up model that guides creative and journey decisions. A rolling cadence of targeted experiments that keep models honest and break ties. A habit of writing down what changed, why, and what to watch next.

It is not perfect. Attribution never is. But it turns attribution from a debate into an operating system. You stop arguing about whose number is right and start asking which move pays back faster, which test would narrow the range most, and which partners earned the next dollar.

That is the quiet power of the approach our team at (un)Common Logic tries to bring into rooms every week. Not a miracle model, not a new acronym. Just enough truth to move money with confidence, again and again, while the world keeps changing around you.