Executive AI Dinner hosted by Kumo - Austin, April 8

Register here
5Counterfactual · Attribution

Marketing Attribution

How much incremental revenue does each marketing touchpoint generate?

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.

By submitting, you accept the Terms and Privacy Policy.

Loved by data scientists, ML engineers & CXOs at

Catalina Logo

A real-world example

How much incremental revenue does each marketing touchpoint generate?

Marketing teams spend millions across channels but cannot measure true incremental impact. Last-touch attribution over-credits the final click; multi-touch models spread credit arbitrarily. Without causal uplift measurement, budget allocation is guesswork — high-performing channels get under-funded while vanity channels consume budget with no real conversion lift. CMOs need to know: 'If we turned off paid search, how many conversions would we actually lose?'

Quick answer

Marketing attribution measures how much incremental revenue each touchpoint actually generates. Counterfactual models compare predicted conversion probability with and without each channel, separating true causal impact from coincidental correlation. This reveals that paid search drives 40% incremental lift for new users but only 3% for existing customers who would have converted anyway.

Approaches compared

4 ways to solve this problem

1. Last-Touch Attribution

Credit 100% of the conversion to the last touchpoint before purchase. The default in Google Analytics, most ad platforms, and CRM reporting.

Best for

Quick reporting when you need a simple answer. Universally available, no setup required.

Watch out for

Massively over-credits bottom-funnel channels (branded search, retargeting) and under-credits awareness channels. A customer who saw 5 ads and then searched your brand name gets 100% credit to branded search. This leads to systematic over-investment in channels that take credit but do not create demand.

2. Multi-Touch Attribution (MTA)

Distribute credit across all touchpoints using rules (linear, time-decay, U-shaped, W-shaped). More nuanced than last-touch but still based on arbitrary weights.

Best for

Teams that want a fairer view of the full funnel without building custom models. A reasonable improvement over last-touch.

Watch out for

The weight distribution is arbitrary. A U-shaped model gives 40% to first touch and 40% to last touch because someone chose those numbers, not because the data supports them. No MTA model answers the causal question: 'Would this conversion have happened without this touchpoint?'

3. Marketing Mix Modeling (MMM)

Aggregate-level regression that correlates total channel spend with total conversions over time. Used by brands with large offline media budgets (TV, radio, out-of-home).

Best for

Measuring the impact of offline channels that cannot be tracked at the individual level. Good for portfolio-level budget decisions.

Watch out for

Operates at the macro level. Cannot measure individual touchpoint impact or distinguish between user segments. Requires 2+ years of data and is typically updated quarterly. Useless for tactical campaign optimization.

4. KumoRFM (Counterfactual Attribution via ASSUMING PQL)

Uses the ASSUMING clause to predict conversion probability with and without each touchpoint. The difference is the true causal uplift. Operates at the individual user level across the full relational graph of users, touchpoints, and conversions.

Best for

Marketing teams that need to measure true incremental impact at the campaign and channel level. Answers the causal question that MTA and MMM cannot.

Watch out for

Requires user-level touchpoint data with timestamps. If your data only has channel-level aggregates without user IDs, you need MMM instead. Also requires historical periods with some variation in channel exposure (not every user saw every ad).

Key metric: Counterfactual attribution reveals that 30-40% of attributed conversions are non-incremental. SAP SALT benchmark: 91% accuracy for relational counterfactual models vs 75% for single-table approaches.

Why relational data changes the answer

User U003 is a new user with no prior brand awareness. Paid search was their first-ever touchpoint. The counterfactual model predicts a 71% conversion probability with paid search and 31% without, a +40% uplift. This is a genuinely incremental conversion. Now consider User U004, a high-value existing customer. Their conversion probability is 91% with paid search and 88% without, only +3% uplift. Paid search gets last-touch credit for both conversions, but the true incremental value is 13x higher for U003.

A flat attribution model cannot distinguish these cases because it does not model the counterfactual scenario. The relational graph enables this by connecting users to their full touchpoint history, segment behavior, and conversion patterns of similar users who did and did not receive each touchpoint. The GNN learns that new users with no organic touchpoints have high paid search uplift, while high-value returning users have near-zero uplift because they would convert through direct or organic channels regardless. On the SAP SALT benchmark, multi-table relational models achieve 91% accuracy vs 75% for single-table models. For attribution specifically, the relational advantage is that the graph naturally encodes the heterogeneity of treatment effects across user segments.

Last-touch attribution is like giving the waiter credit for your restaurant experience because they were the last person you interacted with. Counterfactual attribution asks: what if there was no waiter? You would still eat, but the experience would be worse. What if there was no chef? You would not eat at all. The counterfactual reveals who actually created the outcome vs. who was just present when it happened.

How KumoRFM solves this

Relational intelligence for smarter acquisition

Kumo uses counterfactual PQL queries to answer causal questions. The ASSUMING clause simulates a world where a specific touchpoint exists, and the model compares this to the baseline (without that touchpoint). The difference in predicted conversion probability is the true incremental uplift for that channel. This is not correlation-based attribution — it is learned counterfactual reasoning over the full relational graph of users, touchpoints, and conversions.

From data to predictions

See the full pipeline in action

Connect your tables, write a PQL query, and get predictions with built-in explainability — all in minutes, not months.

1

Your data

The relational tables Kumo learns from

USERS

user_idsegmentsignup_date
U001high_value2025-09-01
U002mid_value2025-09-15
U003new_user2025-10-01
U004high_value2025-10-10

TOUCHPOINTS

touch_iduser_idchannelcampaigntimestamp
T01U001paid_searchbrand_q42025-10-20
T02U001emailnurture_series2025-10-22
T03U002organic2025-10-18
T04U003paid_searchprospecting2025-10-25
T05U004webinardemo_day2025-10-28

CONVERSIONS

conversion_iduser_idrevenuetimestamp
CV01U001$8,4002025-11-01
CV02U004$12,2002025-11-05
2

Write your PQL query

Describe what to predict in 2–3 lines — Kumo handles the rest

PQL
PREDICT COUNT(CONVERSIONS.*, 0, 14, days) > 0
FOR EACH USERS.USER_ID
ASSUMING COUNT(TOUCHPOINTS.*
    WHERE TOUCHPOINTS.CHANNEL = 'paid_search',
    0, 1, days) > 0
3

Prediction output

Every entity gets a score, updated continuously

USER_IDTrue_PROB (with)True_PROB (without)UPLIFT
U0010.870.62+0.25
U0020.340.29+0.05
U0030.710.31+0.40
U0040.910.88+0.03
4

Understand why

Every prediction includes feature attributions — no black boxes

User U003 — new_user segment

Predicted: +40% uplift from paid_search

Top contributing features

New user with no prior brand awareness

new_user

35% attribution

Paid search was first-ever touchpoint

first touch

28% attribution

Campaign — prospecting (top-of-funnel)

prospecting

17% attribution

No organic or referral touchpoints present

0 organic

12% attribution

Similar new users converted 3x more with paid search

3x lift

8% attribution

Feature attributions are computed automatically for every prediction. No separate tooling required. Learn more about Kumo explainability

Frequently asked questions

Common questions about marketing attribution

What is the difference between attribution and incrementality testing?

Attribution assigns credit for conversions that already happened. Incrementality testing (via A/B holdout experiments) measures whether a channel actually causes conversions. Counterfactual models bridge the gap: they estimate incrementality from observational data without requiring costly holdout experiments, updating continuously as new data arrives.

How much budget is wasted due to bad attribution?

Research consistently shows 30-40% of digital ad spend goes to channels and campaigns that would have converted anyway. Branded search and retargeting are the worst offenders. Counterfactual attribution reveals this waste by measuring the actual causal lift, enabling reallocation to truly incremental channels.

Can counterfactual attribution measure offline channels?

Yes, as long as you have user-level exposure data. For channels like direct mail, events, or sales outreach where you know which users were treated, the ASSUMING clause can estimate the counterfactual. For true mass-media channels (TV, billboard) where individual exposure is unknown, MMM is the better approach.

How does counterfactual attribution handle multi-touch journeys?

The model evaluates each touchpoint's incremental contribution by simulating removal. For a user who saw email, then display, then paid search, the model predicts conversion probability after removing each touchpoint one at a time. Touchpoints that increase conversion probability significantly get high attribution; those that do not change the outcome get low attribution.

Bottom line: Counterfactual attribution reveals that paid search drives 40% incremental lift for new users but only 3% for existing high-value customers — enabling precise budget reallocation that can save $2M+ annually in wasted ad spend.

Topics covered

marketing attribution AIcounterfactual attributionincremental lift measurementmulti-touch attributionchannel uplift predictioncausal attribution modelKumoRFMrelational deep learningmarketing mix modelingmedia attributionROAS optimization

One Platform. One Model. Infinite Predictions.

KumoRFM

Relational Foundation Model

Turn structured relational data into predictions in seconds. KumoRFM delivers zero-shot predictions that rival months of traditional data science. No training, feature engineering, or infrastructure required. Just connect your data and start predicting.

For critical use cases, fine-tune KumoRFM on your data using the Kumo platform and Research Agent for 30%+ higher accuracy than traditional models.

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.