Artwork for podcast eCommerce Evolution
Attribution is Broken: Understanding MTAs, MMMs, and Incrementality
Episode 3118th May 2025 • eCommerce Evolution • Brett Curry
00:00:00 00:47:30

Share Episode

Shownotes

In this insightful episode of the E-commerce Evolution Podcast, host Brett Curry sits down with Tom Leonard (https://www.linkedin.com/in/thomasbleonard), a fractional marketing leader who specializes in operationalizing Media Mix Modeling and incrementality testing. They dive deep into the often confusing world of marketing measurement. Tom and Brett will debunk myths about attribution and we reveal what truly drives customer acquisition. 

For ecommerce brands struggling to understand where their marketing dollars are actually working, this conversation offers practical insights on how to move beyond misleading platform metrics.

Sponsored by OMG Commerce - go to (https://www.omgcommerce.com/contact) and request your FREE strategy session today!

Chapters: 

(00:00) Introducing Tom & Marketing Measurement

(06:30) Understanding Multi-Touch Attribution (MTA)

(12:22) The Case for Incrementality Testing

(22:20) Exploring Media Mix Modeling (MMM)

(27:30) Navigating Budget Cuts and Marketing Spend

(32:17) Understanding Incrementality Vs. Attribution

(35:45) The Importance of Cost Per Incremental

(40:16) How to Get Started with MMM

(44:09) Final Thoughts

Connect With Brett: 


Relevant Links:

__


Past guests on eCommerce Evolution include Ezra Firestone, Steve Chou, Drew Sanocki, Jacques Spitzer, Jeremy Horowitz, Ryan Moran, Sean Frank, Andrew Youderian, Ryan McKenzie, Joseph Wilkins, Cody Wittick, Miki Agrawal, Justin Brooke, Nish Samantray, Kurt Elster, John Parkes, Chris Mercer, Rabah Rahil, Bear Handlon, Trevor Crump, Frederick Vallaeys, Preston Rutherford, Anthony Mink, Bill D’Allessandro, Bryan Porter and more

Transcripts

Speaker:

How much is media contributing relative

to customer base is a really nice place

Speaker:

to start.

Speaker:

And the benefit of running

incrementality and media mix modeling is

Speaker:

informing the model with

some of that causal data.

Speaker:

Well, hello and welcome to another edition

of the E-Commerce Evolution podcast.

Speaker:

I'm your host, Brett

Curry, CEO of OMG Commerce.

Speaker:

And today we have got

a doozy of an episode.

Speaker:

We're talking about the three

horsemen of measuring your

Speaker:

marketing effectiveness. We're

talking MTAs Multitouch attribution.

Speaker:

We're talking M'S. Media mixed

modeling. We're talking incrementality.

Speaker:

It's going to be nerdy,

Speaker:

but I also promise you it's going to

be practical and it will make you more

Speaker:

money. And so we'll hopefully

make it fun as well.

Speaker:

And so my guest today is Tom Leonard.

Speaker:

We are LinkedIn friends first.

Speaker:

So I saw Tom on LinkedIn posting about

incrementality, talking about MMM,

Speaker:

throwing shade on certain tools and stuff

like that on LinkedIn. And I'm like,

Speaker:

this is my type of guy. So I reached

out, we had a call, and then we're like,

Speaker:

Hey, we got to record a podcast.

Speaker:

Let's create some insights

for people on the pod.

Speaker:

And so Tom is a fractional

marketing leader.

Speaker:

He's operationalizing MMM

and incrementality testing,

Speaker:

and I'm delighted that he's my guest

today. So Tom, with that intro,

Speaker:

how's it going? And welcome to the show.

Speaker:

Good. Thanks for having me, Brent.

Excited to be here. And yeah,

Speaker:

some of my favorite things to talk

through, so excited to do it. Good stuff.

Speaker:

It's good stuff, man. So briefly,

Speaker:

before we dive into the

meat of the content here,

Speaker:

what's your background and

how did you become a guy who's

Speaker:

operationalizing MMS and incrementality?

Speaker:

Yeah. And what does that even mean?

Speaker:

That's a good point.

Speaker:

For sure. Yeah, totally. Yeah.

Speaker:

So spent most of my career thus far on

the agency side at performance agencies.

Speaker:

And I'd say the crux of

how I got to where I'm now,

Speaker:

or I've been reflecting back a little

bit more on the why I have such a passion

Speaker:

for measurement. And I was at

a pretty hardcore DR agency,

Speaker:

and it was right shortly after TRUBY

for Action came out when YouTube was

Speaker:

starting to invest in, DR.

Speaker:

Moved into a new role we had created

with a centralized group of basically

Speaker:

people who had different areas of subject

matter expertise and a few analysts

Speaker:

that ran tests across a

pretty large client base.

Speaker:

And I was our YouTube SME,

Speaker:

and worked with a couple

analysts to run a bunch of tests.

Speaker:

And really it was to evangelize how to,

Speaker:

and is YouTube a platform to drive growth?

And it was really interesting

Speaker:

because I started spending a lot of time

on YouTube and then also connect to TV

Speaker:

and broader programmatic video.

And it was this really interesting,

Speaker:

for me, the biggest learning was less

about how to make YouTube as effective as

Speaker:

possible,

Speaker:

but more how to help brands think about

demand creation as opposed to just

Speaker:

demand capture. And frankly,

Speaker:

the difficulty of getting brands

to leverage YouTube relative

Speaker:

to connected tv,

Speaker:

because YouTube sat so close to Google

ads and therefore last click attribution

Speaker:

and see tv, you couldn't click

and was sexier in a deck.

Speaker:

And it was just this sort

of recognition of the

Speaker:

irrational kind of human behavior just

in any sort of industry or any thing

Speaker:

in life.

Speaker:

But it sort of helped frame up this

idea of you really have to do more than

Speaker:

just, I don't know,

Speaker:

represent logic or rational arguments.

You really have to also

Speaker:

bring the easy to understand

clear data. And that's,

Speaker:

I think what draws me to incrementality

testing specifically and why

Speaker:

that's sort of the backbone

of a lot of what I do now.

Speaker:

And I think I use the word

operationalizing, NMM and

incrementality testing.

Speaker:

And really what I mean by that is a lot

of people will run medium mix models or

Speaker:

run incrementality tests,

Speaker:

but oftentimes they'll sit in a slide

or in a report to be shown once,

Speaker:

but never to be looked at again.

Speaker:

And so what I'm really trying to do

with brands now is how do you build a

Speaker:

framework and a repeatable methodology

to get insights from tests,

Speaker:

but not just leave them as

insights but to take action?

Speaker:

Because the only way that you create

value from any of these sort of testing

Speaker:

methodologies and measurement

methodologies is by

acting on the insights.

Speaker:

And so that's sort of what I mean by my

funky little headline of those words.

Speaker:

Yeah, it's so good, man.

Speaker:

And it's one of those things where data

really doesn't matter if you don't take

Speaker:

the right actions from it.

And what's so interesting,

Speaker:

and our paths are similar in that

I got my start in actually TV and

Speaker:

radio and doing traditional media, and

then I got into SEO and paid search,

Speaker:

but I loved video. Video was my

thing, but I love paid search as well.

Speaker:

And then when TrueView and TrueView

for Action came out, I was like, whoa,

Speaker:

these are all my world's colliding.

Speaker:

This is.

Speaker:

Video and there's some search components,

Speaker:

at least some search intent involved

there. And it's direct response.

Speaker:

I've always been a direct response guy.

Speaker:

I believe that marketing

should drive an outcome, right?

Speaker:

Advertising should drive

a measurable outcome,

Speaker:

and that should be measured in terms

of new customers and profitable new

Speaker:

customer acquisition. And

what's really interesting, Tom,

Speaker:

and I think this kind of feeds into

the conversation we're having today.

Speaker:

There was a period of time, so I

grew up reading some of the classics.

Speaker:

So David Ogilvy of course, but John

Cap's tested advertising methods,

Speaker:

Claude Hopkins Scientific Advertising.

Speaker:

And they would do things like they would

run and add in a newspaper or magazine

Speaker:

and people would clip a

coupon and bring it in,

Speaker:

or they would call a certain number and

they would track it and they would have

Speaker:

codes and stuff.

Speaker:

And I remember thinking once I got

into e-commerce, I was like, oh man,

Speaker:

we've got so many tools. The world is

so clear now we have every piece of

Speaker:

data at our disposal.

Speaker:

And now the more I've gotten into it

and the more I've matured, I'm like,

Speaker:

we've got more data. But I don't

know that we've got more insights,

Speaker:

and I don't know that we've

got any more clarity. In fact,

Speaker:

there's maybe more confusion.

Speaker:

And I think it goes back to

what you said a minute ago,

Speaker:

this idea of demand generation

versus demand capture.

Speaker:

We're really good at measuring channels

and campaigns that are demand capture,

Speaker:

meaning they're capturing

demand that's already out there.

Speaker:

That's harder to measure

the demand generation,

Speaker:

which is usually where the magic happens.

Speaker:

And so super excited to dive in here.

Speaker:

I think what might be useful

is let's talk about what

Speaker:

are these kind of three horsemen that

I laid out there, MTAs, multitouch,

Speaker:

attribution, and incrementality.

So let's start with MTAs first.

Speaker:

So Multitouch attribution tools,

Speaker:

what are they and what

is your take on them?

Speaker:

Yeah, big question. Great

question. Yeah, I mean,

Speaker:

MTA been around for a while,

Speaker:

different flavors and ways

of trying to make it work,

Speaker:

especially as so much has changed

in privacy and the tech and tracking

Speaker:

landscape.

Speaker:

But ultimately the goal is to try

to give fractional credit to all the

Speaker:

touchpoints along a customer journey with

a recognition that the last touchpoint

Speaker:

click or last impression is

ultimately not what drove that person

Speaker:

to purchase.

Speaker:

That may be the last or the only thing

that you might see in something like

Speaker:

Google Analytics or your analytics suite.

Speaker:

But there's this general recognition

that that is not what drove the purchase.

Speaker:

So MTA, the kind of promise, which I

ultimately think is a failed promise,

Speaker:

is whether all the different touch

touchpoint and then how can you

Speaker:

value those differently. So

maybe you use first touch,

Speaker:

maybe you use even distribution. The

idea of data-driven attribution was the

Speaker:

holy rail or the Promise many years ago,

Speaker:

and I guess still to a

degree for some is like,

Speaker:

how do you know this channel was more

additive or more necessary and therefore

Speaker:

should get more credit than that channel?

Speaker:

Which I think makes a

ton of sense in promise.

Speaker:

I think in reality it's really hard

and I would argue impossible to do,

Speaker:

especially as a lot of the ability to

track users at a one-to-one level degrades

Speaker:

generally my perspective,

I'm very bearish on MTA,

Speaker:

so that'll probably come

through pretty strongly.

Speaker:

But I guess I don't think the toothpaste

is going back in the tube in terms of

Speaker:

the ability to track a customer across

all these different touchpoints,

Speaker:

especially as the ability to

track through or impression based

Speaker:

touchpoint erodes. And then you

really get reliant on clicks,

Speaker:

which I think then leads to a lot of

all the issues that just last click in

Speaker:

general has.

Speaker:

So I think it's really hard to

make a compelling case for MTA.

Speaker:

I've seen too many brands,

Speaker:

especially trying to

build MTA tools internally

Speaker:

and just be a huge time and resource

suck. And then when you ask to compare,

Speaker:

show the multi-touch view versus

last click, it's like, I don't know,

Speaker:

80 or 90% only had one touch

point anyways, that's all

that MTA model could see.

Speaker:

So is it really that much

more useful than last click?

Speaker:

It's sort of multi-touch when that can

be measured, but usually it can't be.

Speaker:

Yeah, and It never really answers

the causality question either,

Speaker:

which we'll get to when we

talk about incrementality.

Speaker:

And I always kind of tell this,

Speaker:

I think the short story of why MT A

isn't really viable anymore as all the

Speaker:

tracking and privacy changes.

Speaker:

But I think the slightly longer story

is the kind of recognition that just

Speaker:

because an ad was shown or a

click occurred doesn't mean that

Speaker:

that medium was needed or

that channel was needed.

Speaker:

It doesn't answer the causal question,

Speaker:

what would've happened

without this ad running?

Speaker:

Did somebody just happen to use multiple

touchpoints as navigation or was it

Speaker:

more convenient to click on one of

these ads that happened to be served?

Speaker:

But if you're not comparing that to some

sort of control group to really hard

Speaker:

to assign causality to the fact

that there just was a touchpoint.

Speaker:

Yeah, it is so good. And it's one of

those things where I remember again,

Speaker:

early on,

Speaker:

you would look inside of Google ads or

you look inside of Meta or was back when

Speaker:

it was Facebook only, and you

were like, the data's here.

Speaker:

I see row ads and I see clicks and

I see performance and all that.

Speaker:

Then you realize, well, wait a

minute, this isn't fully accurate.

Speaker:

If I add the two together,

that's double my total revenue,

Speaker:

so I can't just rely on

what's in the platform.

Speaker:

And that got worse as I was 14 was

introduced and other privacy changes were

Speaker:

made. But then MTA came

along and it's like, oh,

Speaker:

finally we're going to get to see the

full picture. It's going to decipher,

Speaker:

decode the shopping journey,

Speaker:

and we're going to finally see with a

keen eye in perfection exactly what caused

Speaker:

this ad or what caused this purchase

to happen. And then we finally realized

Speaker:

MTA is maybe just a third

option. It's like, okay,

Speaker:

Google's imperfect, Meta's

data's imperfect, and then mt A,

Speaker:

it's just imperfect too.

Speaker:

So now we just got three imperfect

things to look at and make

Speaker:

decisions from.

Speaker:

And in some ways it leads to more

confusion than it leads to clarity.

Speaker:

And now I don't want to wholesale discard

Speaker:

MTAs because I do believe there's some

helpful insights that can be gained

Speaker:

there,

Speaker:

but it's incomplete

and incomplete at best.

Speaker:

And one of the best analogies I've heard,

and this actually comes from Ben Ter,

Speaker:

who's also a LinkedIn friend,

but I met him in person as well,

Speaker:

but he talks about this analogy of, Hey,

Speaker:

if we're trying to measure what

caused people to watch this

Speaker:

movie at our movie theater,

Speaker:

and we look at all these

results and 30% say they saw a

Speaker:

billboard for our movies,

20% say they saw a TV ad,

Speaker:

but you know what? A hundred percent

say they saw the poster on the

Speaker:

door. So we're like,

let's just cut everything.

Speaker:

Let's just do the poster at the door

and that's it. And you're like, well,

Speaker:

wait a minute. Everybody saw it.

Everybody was walking in the door.

Speaker:

But the movie poster is not

what caused someone to purchase.

Speaker:

It was the billboard and the TV

and some of the other things,

Speaker:

word of mouth and other things

that caused them to come in.

Speaker:

And so this idea of causality,

super, super valuable.

Speaker:

So that really leads us to incrementality.

So talk about incrementality.

Speaker:

What is it and why are you on

a quest to operationalize it?

Speaker:

Yeah, it's really the best way,

Speaker:

if not the only way to

establish that a causal

Speaker:

portion that we've been talking about.

It has a distinct control group,

Speaker:

so it has a counterfactual,

Speaker:

it has what would've happened

without this intervention,

Speaker:

whatever that intervention is.

Speaker:

And there's a handful of ways to derive

that counterfactual that control.

Speaker:

The most common would be geographic

based. So like a match market test.

Speaker:

I've got this market over here that

historically has behaved similarly to this

Speaker:

market over here. I can

see that in an AA test,

Speaker:

the lines sort of move similar

to one another. They're not,

Speaker:

if they're influenced by outside

factors, they're influenced.

Speaker:

In what's an AA test for

those who don't know.

Speaker:

Before an intervention happens.

Speaker:

So just over time are those lines

essentially moving together?

Speaker:

Are external factors or stimuli equally

impacting both sides of that test

Speaker:

so that you can feel confident that

when you do intervene and it becomes

Speaker:

comparing A to B,

Speaker:

the delta is what was a

result of that intervention.

Speaker:

So oftentimes it's my Atlanta

Speaker:

and I don't know Memphis,

Speaker:

maybe some other midsize city that

you've done this market matching for.

Speaker:

Historically, they both

look like this on a line,

Speaker:

all of a sudden you turn off

ads on Facebook in Atlanta,

Speaker:

what happens to your top line that

Delta is what was attributed or

Speaker:

should be attributed to

advertising in Atlanta.

Speaker:

Whereas the flip side of that would be

attribution would say basically anything

Speaker:

that was attributed to that could

be attributed to that would really,

Speaker:

it should just be the gap between a

world where that ad does not exist

Speaker:

compared to a world where that ad

does exist. We can't take credit for

Speaker:

everything.

Speaker:

We can only take credit for as much

above and beyond what would've happened

Speaker:

anyways. And so that's the

basis of incrementality testing.

Speaker:

There's other ways to do it.

Speaker:

If you use a Facebook or Google

conversion lift study because they own

Speaker:

that auction or anybody

that owns an auction,

Speaker:

they can do that hold out

for you at a user level.

Speaker:

They can track all of those users

regardless of if you serve an ad.

Speaker:

Good examples are maybe easier to

describe in a first party data capacity.

Speaker:

If you're running email, you may blast

all of your customers and say, Hey,

Speaker:

I sent an email to all my

customers and this many purchased.

Speaker:

They went back to the website or

clicked it. But if you just said, Hey,

Speaker:

I'm going to serve just to odd

number of customer IDs and not to

Speaker:

even number customer IDs,

I can then just compare,

Speaker:

forget about who clicked on ads,

Speaker:

who did anything.

I'm just going to look at my backend.

Speaker:

I know I exposed these users,

but not these users 50 50 split.

Speaker:

They've historically kind

of done the same thing.

Speaker:

All I did was even an odd and just

measuring the difference between those two

Speaker:

groups.

Speaker:

So really any way that you can

establish a true control that

Speaker:

passes that AA test. So

before you intervene, do they

continue to look similar?

Speaker:

Are they influenced at the same rate so

that you can feel confident that when

Speaker:

you do intervene with new

media, retracting media,

Speaker:

some new sort of test that you are

confidently comparing to what would've

Speaker:

happened in a world

without that intervention?

Speaker:

Yeah, yeah.

Speaker:

It's applying the scientific

method with some rigor behind

Speaker:

what happens when I turn this channel on,

Speaker:

or what happens when I

turn this channel off?

Speaker:

What is the actual impact of this channel?

Speaker:

And what's interesting is I

remember back in my early days

Speaker:

of being in the advertising world,

Speaker:

this was when online stuff was

just getting kind of warmed up.

Speaker:

I was talking to this furniture store

owner and I'm like, Hey, what do you do?

Speaker:

Do you invest in radio ads?

Do tv, do you do newspaper?

Speaker:

And so as I went through Themm like,

Hey, do you do radio ads? And he is like,

Speaker:

yeah, I mean, yeah, I sort of do.

And I'm like, newspaper's like, yeah,

Speaker:

there's a big sale, something will

happen. I'm like, well, what about tv?

Speaker:

And he said, yes. And his

eyes lit up and he is like,

Speaker:

when I run TV ads, I feel

it. People walk in the door,

Speaker:

it happens. And I remember early on

in my online career thinking, man,

Speaker:

that was so unsophisticated. Did

that guy really know what's going on?

Speaker:

But now looking back, I'm like,

yeah, that's maybe all that matters.

Speaker:

That is incrementality in a real loose

easy just to observe with your eyes think

Speaker:

because you had one. Totally.

Speaker:

Which I think people

take for granted. Yeah.

Speaker:

They do.

Speaker:

Yeah.

Speaker:

That's not exciting. That's not

like, where's all your data?

Speaker:

It's in my cash register.

That's where all the data.

Speaker:

Is, especially for smaller brands,

Speaker:

when you have the ability

to feel if something's

Speaker:

working or not working,

Speaker:

if you double spend in something that

you think is working really well because

Speaker:

attribution says it's working really well,

Speaker:

and all of a sudden

your cash just doubles,

Speaker:

even though your attributed number

scales linearly, something has to give,

Speaker:

right?

Speaker:

And what has to give is it wasn't really

causing any additional top line growth.

Speaker:

It was just really good at

getting the attributed credit.

Speaker:

So I think the feeling

it in the p and l is

Speaker:

definitely overlooked.

Speaker:

It's valid, and it is overlooked

though. You're a hundred percent,

Speaker:

especially now that we have

so many tools at our disposal.

Speaker:

And I think another way to look at

this, and look, I'm a Google guy,

Speaker:

YouTube and Google is kind of where

I really got my start in online.

Speaker:

Marketing.

Speaker:

But listen, branded search is a

perfect example here. What happens,

Speaker:

we see this all the time.

Speaker:

What happens if you turn branded

search completely off? Now, I believe,

Speaker:

and this is top of front of the podcast,

Speaker:

there are strategic ways to use branded

search and there's ways to run it and

Speaker:

not waste money, but a lot of people

could shut it off and nothing happens,

Speaker:

nothing. Maybe sales get in a little bit,

Speaker:

but you take meta meta's really working

and you shut it off and you feel it.

Speaker:

Sales go down and that's

an incrementality.

Speaker:

Same is true for YouTube if you're doing

YouTube the right way. And so yeah,

Speaker:

I really like this. And one

kind of anecdote here to share,

Speaker:

we just did a test with Arctic,

Arctic coolers, Yeti competitor,

Speaker:

my favorite cooler, my favorite drinkware

as well. And so they wanted to see,

Speaker:

Hey, can YouTube drive an incremental

lift at Walmart? So they had just

Speaker:

gotten into most Walmart

stores, coast to coast.

Speaker:

So we did exactly what you laid out

there. We had a 19 test markets,

Speaker:

19 matched control markets.

So similar markets.

Speaker:

So think like a Denver and a

Kansas City or the example,

Speaker:

use Atlanta and whatever else

that's kind of comparable. And hey,

Speaker:

let's run YouTube in one

and not in the other.

Speaker:

And let's measure then the

growth in Walmart sales,

Speaker:

and let's do a comparison

between the two in Walmart sales.

Speaker:

And it was remarkable. It

was about an eight week test.

Speaker:

We had three test regions, so 19

markets, but three test regions,

Speaker:

test region. One, we saw an average

of 12% lift in Walmart sales.

Speaker:

The test region two was like 15% lift.

Speaker:

And then our final test

region was 25% lift.

Speaker:

And there were some standouts,

Speaker:

like Oklahoma City was up 40% and Salt

Lake City was up 48%. But it was one of

Speaker:

those things where, okay, now we

look at that and we can say, okay,

Speaker:

YouTube had a big impact. And

what's also interesting, Tom,

Speaker:

is we just ran the YouTube portion at OMG.

Speaker:

They also did a connected TV test

in other markets, not related,

Speaker:

didn't see a lift, didn't

see a measurable lift.

Speaker:

And so it could be lots of

that was not to throw shade on

Speaker:

CTVI like CTV,

Speaker:

so maybe they just did a wrong or

wrong creatives or who knows what.

Speaker:

But it's one of those things

where it's like, okay,

Speaker:

if you do this the right way,

you should see an impact.

Speaker:

And I think touching on the

piece that I didn't mention,

Speaker:

the other beauty or value of

incrementality testing relative to

Speaker:

attribution or mt a is the ability

to see beyond your.com to be able to

Speaker:

see what's happening on third parties

like Amazon, what's happening in store.

Speaker:

If you get that data own an operated

store or if you can get that through

Speaker:

wholesale data, it really simplifies.

Speaker:

There's so much complexity.

And I think that's, again,

Speaker:

one of the rubs that I have

with MTA is all of them,

Speaker:

all of the data you have to

wrangle together to try to

Speaker:

patchwork this kind of story together.

Speaker:

Whereas in incrementality testing,

it's pretty straightforward.

Speaker:

It's what did I spend and how

did I run that spend in these by

Speaker:

market by day or by week, and what

was my sales? What were my sales?

Speaker:

What were my new customers or whatever

metric I'd want to look at with that same

Speaker:

granularity and same dimension.

Speaker:

And that's really it because you're

really just trying to understand the

Speaker:

relationship that calls the

relationship between spend and outcomes,

Speaker:

all that kind of muddy middle

in the middle, trying to

get it at the user level,

Speaker:

which again, not going back into

the tube really simplifies things.

Speaker:

Yeah, it does.

Speaker:

And another thing that was

kind of interesting that

came a light doing this test

Speaker:

for Arctic is all of the ads we

tagged with available at Walmart,

Speaker:

shop at Walmart, find on the

shelves and Walmart, whatever.

Speaker:

We measured everything

though in those markets.

Speaker:

So you could look at Walmart sales,

online sales, so the.com and Amazon.

Speaker:

And what's interesting is the

push to Walmart really worked.

Speaker:

It's a reminder of what you ask someone

to do in an ad is what they're going to

Speaker:

lean towards. Because

in some of the markets,

Speaker:

we didn't see that much of an online lift.

Speaker:

We saw some clicks and stuff like

that, but the Lyft was at Walmart.

Speaker:

But we also saw a pretty

strong lift at Amazon as well,

Speaker:

because I think that just speaks to,

Speaker:

there's some people that are just going

to buy everything from Amazon right

Speaker:

there, tell 'em to go online value pro

proposition. Is it on Amazon? Yeah, yeah.

Speaker:

Yeah. Here in a day or two, it's hard.

Speaker:

To beat, dude. It's hard to beat

same price in a couple days.

Speaker:

I don't have to leave my house. But

yeah, really, really interesting.

Speaker:

And so we'll circle

back to that of course,

Speaker:

but let's talk about then

MMM or media mix modeling.

Speaker:

What is that? How are you using that?

Speaker:

And then how does that kind of relate to

incrementality testing? Because again,

Speaker:

going back to your tagline, Tom, you

did not say operationalizing NTAs.

Speaker:

You said operationalizing m

and ms and incrementality.

Speaker:

So what is MM and how does

that pair with incrementality?

Speaker:

Yeah,

Speaker:

basically a big correlation exercise

trying to suss out without a true kind of

Speaker:

holdout group,

Speaker:

what is the impact and contribution of

each media channel and also what would

Speaker:

happen without media.

Speaker:

So trying to suss out a lot of the

same questions as incrementality,

Speaker:

but basically using correlation as

opposed to having a true holdout group.

Speaker:

So basically,

Speaker:

and I'm sure all the hardcore MMM people

and data scientists will thumbs down

Speaker:

this or whatever you can do to podcast,

but hey, in this period of time,

Speaker:

sales went up and nothing could really

explain that other than the fact that

Speaker:

TikTok spend went up and essentially

doing that at a mass scale over longer

Speaker:

periods of time trying to take into

account anything that could explain that.

Speaker:

So you'll always kind of flag it with

these are promotions that happen,

Speaker:

it should because you're going to give

a model at least like two years worth of

Speaker:

data or two years worth of data,

Speaker:

it'll bring in seasonality and try to

understand those sort of trends. So it's

Speaker:

trying to pull out if not

seasonality, if not promotions,

Speaker:

if not some other things

that we are flagging.

Speaker:

And it wasn't price reductions,

it wasn't all these pieces,

Speaker:

what was happening in media

that could explain that change.

Speaker:

And so that's ultimately

what MMM is doing.

Speaker:

It's a big correlation exercise,

Speaker:

figuring out roughly what is the channel

contribution to a top line revenue or

Speaker:

order number and what's really important.

Speaker:

I think the nicest part or the best

first step with M is trying to get an

Speaker:

understanding of a base,

Speaker:

which is what it's going to be called or

intercept what without the presence of

Speaker:

ads,

Speaker:

does this model think that my sales would

be such that I can then calculate not

Speaker:

a total CAC of just looking at

total new customers divided by cost,

Speaker:

but incremental to media

or remove base from

Speaker:

that equation,

Speaker:

how many conversions were contributed

because of media as this model sees,

Speaker:

which no model is going to be perfect,

Speaker:

no measurement method

is going to be perfect,

Speaker:

but it's a really nice

place to start to say,

Speaker:

I knew I couldn't account all

new customers to advertising,

Speaker:

but what's a good number to use or

to start with? Well, it looks like,

Speaker:

and this will depend on the maturity of

the brand, but a really mature brand,

Speaker:

I mean super mature brand,

Speaker:

the big CPGs might be like 99% base

smaller brand might be something

Speaker:

like 50% because you've got

this word of mouth flywheel,

Speaker:

you've got product market fit,

Speaker:

but trying to get an understanding of how

much is media contributing relative to

Speaker:

customer base is a really

nice place to start.

Speaker:

And the benefit of running

incrementality and media mix modeling is

Speaker:

informing the model with

some of that causal data.

Speaker:

You see that a lot and there's a

really powerful feature of media mix

Speaker:

modeling is saying, Hey, yes,

that's a correlation exercise,

Speaker:

can't pull everything out,

Speaker:

but let me inform the model or at least

restrict the priors it can use or the

Speaker:

coefficient, whatever

you want to call 'em,

Speaker:

what it's searching for to try to find

a fit in this model and say, well,

Speaker:

I did a hold out test. I know

you don't have the causal data,

Speaker:

but we ran this in this channel and that

channel and helping that restrict the

Speaker:

model and giving it data that it can't

have without that human intervention can

Speaker:

be a really powerful flywheel.

Speaker:

So using your incrementality test data,

Speaker:

feeding that back into your MMM

model to make it more accurate and

Speaker:

more causal and make that correlation.

Speaker:

Stronger.

Speaker:

Because the two things that are really

like you're really trying to get,

Speaker:

but you don't get with Multi-Tech

attribution or attribution in general.

Speaker:

And you do get with the combination of

media mix modeling and incrementality

Speaker:

testing is the incremental impact,

Speaker:

the causal impact of what

would've happened without

the presence of ads as well

Speaker:

as the diminishing returns curve,

Speaker:

which we know can be really

powerful and important too,

Speaker:

is what has happened over time as I

spend in that sort of a feature of big

Speaker:

feature of media mix modeling

is understanding where

are you on a diminishing

Speaker:

returns curve? Is there

if I keep spending more,

Speaker:

I know it's not going to scale linearly,

Speaker:

but are there channels

that diminish faster?

Speaker:

Is there more headroom in other channels?

Speaker:

And it really becomes this

true optimization game of

where do I put the next

Speaker:

dollar? Ultimately the

question that every marketer,

Speaker:

every finance team is

trying to answer is, Hey,

Speaker:

if I find $20,000 into couch

cushions, where do I put it?

Speaker:

And if I need to give back $20,000,

where do I pull from to have.

Speaker:

I want to hang out at your house and

look at your couch cushions and find 20

Speaker:

grand? That's.

Speaker:

Great. Yeah, it's easy to

give it back, but yeah, right.

Speaker:

We're trying to figure out what is going

to be the least impactful if I have to

Speaker:

give the money back and cut budgets

and where is it going to be the most

Speaker:

impactful if I have another $20,000?

Speaker:

Because the answer is not going to be

found in what has the highest or the

Speaker:

lowest ROAS in an attributed

view. And in fact,

Speaker:

that can have the complete

opposite impact that you want.

Speaker:

Yeah, yeah, it's really great.

Speaker:

So I want to actually talk about

that point in a minute where

Speaker:

if you've got cut budgets,

which hey, listen,

Speaker:

there's been some uncertainty even as we

record this, tariffs up, tariffs down,

Speaker:

markets up, market down, whatever

consumer sentiment is all over the place.

Speaker:

So if things get a little bit

tight, what are we going to do?

Speaker:

We can't slash marketing,

we can't slash growth.

Speaker:

I think that sends you

into a death spiral,

Speaker:

but we might have to get pull

back and get more efficient.

Speaker:

And so let's talk about that

actually for a little bit.

Speaker:

So where can you be led astray?

Speaker:

I think you just made a post

on LinkedIn about this, right?

Speaker:

Where you start looking at performance,

which feels like the smart thing to do,

Speaker:

looking at ROAS and whatnot, and

you're like, well, great, well,

Speaker:

let's just cut the lowest ROAS

campaigns and channels. We'll be fine.

Speaker:

How does that lead you astray?

Speaker:

And if you want to talk about your

specific example to help illustrate these

Speaker:

points, that'd be great.

Speaker:

Yeah, totally.

Speaker:

I think the other one you're referring

to is I think branded search,

Speaker:

which we were talking about

earlier. And I love using both a,

Speaker:

because it can be really, if a brand

is spending a lot of money there,

Speaker:

it can be a really great place to go

find those savings without impacting top

Speaker:

line. But also frankly, it's

really easy to understand.

Speaker:

I think most people understand that

up and down the organizational chart

Speaker:

across departments, everybody sort

of understands the idea of, Hey,

Speaker:

if somebody's already

searching for my brand,

Speaker:

do I need to pay to get that

click and that conversion?

Speaker:

And I found that just the fact that

it's easy to understand can be a

Speaker:

really good gateway to incrementality

testing because it's easy to get buy-in.

Speaker:

Everybody understands that idea,

Speaker:

whereas it may be more challenging

to express that idea in

Speaker:

other types of campaigns.

But branded search is a good example,

Speaker:

and the example that you're referring to,

Speaker:

kind of a midsize brand that I was

working with went through that exact

Speaker:

exercise, had to cut budgets.

Speaker:

They looked at up and down the campaigns

they were running. It was like, Hey,

Speaker:

we just got to make the best decision

we can with the best available data.

Speaker:

They were basically running p max

non-branded search and branded search and

Speaker:

p max and branded search where had

the best attributed roas Best CPA

Speaker:

non-brand was really hard to justify in

a lower budget kind of environment based

Speaker:

off the attribution data cut that leaned

a little bit more into branded search

Speaker:

as a percentage of their budget.

And over the next couple months,

Speaker:

new customers in total revenue

was declining despite the

Speaker:

attributed ROAS and CPA

looking even better than ever.

Speaker:

And that's where was brought

in, looked at all these things,

Speaker:

saw the loose correlation to

non-brand and new customer

Speaker:

acquisition and top line,

Speaker:

just the general skepticism that

many have around branded search,

Speaker:

especially in a low

competition environment,

Speaker:

which they were in. There weren't many

competitors in the auction that we

Speaker:

could see in Auction Insights. So yeah,

Speaker:

ran a very blunt instrument

match market test,

Speaker:

which at a brand of that size and for a

branded search I don't think is ever a

Speaker:

bad idea. And yeah, no

impact to branded search.

Speaker:

It was about 20% of their budget,

Speaker:

which was substantial that you

can either make the decision,

Speaker:

I'm going to put that 20% back in

my pocket or save it for a rainy day

Speaker:

or give it to some other

place in the org or say, Hey,

Speaker:

I'm going to redistribute this to

something that I see in correlation

Speaker:

data that might help

drive top line backup.

Speaker:

Let's reinvest that in non-brand as

opposed to keeping it in branded. Again,

Speaker:

complete opposite of what

attribution would say.

Speaker:

And you see that a lot frankly with

branded search is an easy one to pick on.

Speaker:

Same with retargeting,

Speaker:

but really anything that's especially

challenging with the black box

Speaker:

solutions that blend,

Speaker:

and I'm sure we could do a whole talk

show on p max Advantage plus some of the

Speaker:

things that bundled together historically

radically different levels of

Speaker:

incrementality can be a real challenge

when you're then measuring on

Speaker:

attribution. But yeah, a

ranty way of saying yes,

Speaker:

finding areas to cut oftentimes

if you follow the attribution kind

Speaker:

of data can lead to really kind

of impactful in a negative way

Speaker:

business outcomes because the attribution

view just does not take into account

Speaker:

what would've happened

without the presence of those

ads like Incre Ality does.

Speaker:

And so can definitely lead brands

astray as they're looking to cut.

Speaker:

Yeah, really interesting. And yeah,

Speaker:

max notorious for leaning into

remarketing or branded search.

Speaker:

If you're not diligent about that, it

can lean into both of those things.

Speaker:

And so got to be mindful of that.

Speaker:

You also quoted something

that totally ties into this.

Speaker:

It's from a shop talk talk that

you went to shop Talk the show,

Speaker:

and I can't remember who said

it, but if you see high roas,

Speaker:

I know something is wrong and that the

auto targeting is just finding existing

Speaker:

customers. Do you remember actually

who said that and unpack a little bit?

Speaker:

Yeah, I forget his name and I could

look real quick. He worked for.

Speaker:

Mic.

Speaker:

The Post Dan Danone, the big CPG.

Speaker:

Yeah, I just really appreciated

that quote because I

Speaker:

mean always wonder if I live in sort of

a bubble of being super passionate about

Speaker:

incrementality versus attributed metrics,

Speaker:

but that was just really refreshing to

hear because I don't think that's the

Speaker:

natural.

Speaker:

It's not.

Speaker:

Thought in people's.

Speaker:

Head spend more.

Speaker:

But I really think it should

kind of spark some skepticism,

Speaker:

especially when your goal really

is to try to drive new customers.

Speaker:

My first,

Speaker:

especially if you think about both

incrementality in the context of a SC

Speaker:

or pex that's blending retargeting

and prospecting by default

Speaker:

and knowing diminishing returns

Speaker:

are my first dollars, yes, they're

going to be the most effective,

Speaker:

but if they are focused on people that

are already buying from me and my goal in

Speaker:

my head is new customers,

Speaker:

I should be shocked that I can

spend a hundred dollars and drive

Speaker:

this amazing new customer revenue

Speaker:

and not think that something is up or

even over time as I continue to spend

Speaker:

our BS meters should probably

go up a little bit more.

Speaker:

And I don't think they do by default. So

I found that comment really refreshing.

Speaker:

Yeah, I think that

really illustrates that,

Speaker:

right where it's like most of us would

think, oh, ROAS is going up great,

Speaker:

we're printing money.

Speaker:

Whereas maybe you should say BS

detector, something's wrong here.

Speaker:

This campaigns leaning into customers

that we're going to buy anyway.

Speaker:

And I'll give two examples here to

illustrate this a little bit more.

Speaker:

And I'll also, since we've been

picking on branded search so much,

Speaker:

I'll share a couple of ways I

think we should use it. One.

Speaker:

If.

Speaker:

Other competitors are

aggressively bidding on,

Speaker:

just know that if you're not Nike and

you're not Adidas and you're not like Ford

Speaker:

or something, it's not a

lock. If it's a new customer,

Speaker:

they could be swayed by a competitor.

Speaker:

And that's generally how we

like to separate it out is like,

Speaker:

let's have branded search for returning

customers and let's make that crazy

Speaker:

efficient or just turn it off altogether.

Speaker:

If.

Speaker:

It's a new customer, then again,

we want it to be very efficient,

Speaker:

but maybe we want it on because we

don't want our competitor to come in and

Speaker:

swipe us to give and swipe our

customer. And so one example of this,

Speaker:

I did a podcast with Brian Porter,

he's the co-founder of Simple, modern,

Speaker:

great Drinkware brand has become a friend

and they did a study incrementality

Speaker:

study and they found, I'll

get these numbers off,

Speaker:

but it was like branded

search was 10% incremental.

Speaker:

So basically what that means is if it

shows that I got a hundred new customers

Speaker:

from Branded Search,

Speaker:

I probably would've gotten 90 of

those if I had shut it off, right?

Speaker:

Only 10% were incremental.

Speaker:

So then what you would need to do there

is you need a 10 x row as on branded

Speaker:

search for it to even make

sense. If it's below that,

Speaker:

you're completely wasting

money. Pair that with,

Speaker:

and you and I were commenting

on the House analytics, HAUS,

Speaker:

Olivia Corey and team did 190

incrementality studies involving

Speaker:

YouTube and they showed with

tremendous amounts of rigor

Speaker:

that hey,

Speaker:

YouTube is probably 342 times more

Speaker:

incremental, meaning if

you see a one in platform,

Speaker:

it's actually like a 3 42 in

terms of incremental impact.

Speaker:

And so wildly different

between those two. But again,

Speaker:

we're just so drawn to in platform

row as man, we'll just say spin,

Speaker:

spin spend on p max and branded search

when really we should be saying,

Speaker:

let me lean into YouTube or let

me lean into top of funnel meta.

Speaker:

I think both those examples

too are really good examples.

Speaker:

To me it also speaks

though to the importance of

Speaker:

cost per incremental almost being

more important than incremental

Speaker:

percent incremental. And that's something

I always use with branded search.

Speaker:

I think you and I have a very similar

feeling around branded search.

Speaker:

There's definitely a

time and a place for it,

Speaker:

and it's one of those things where

it might not matter that it's 10%

Speaker:

incremental, 10% incremental relative

to what Google's attributing.

Speaker:

If your attributed CPA

is a dollar and now it's

Speaker:

$10,

Speaker:

but your margin when you sell a

product is a thousand dollars like

Speaker:

hammer that all day long,

Speaker:

that cost per incremental is still

extremely profitable and valuable.

Speaker:

And same with the YouTube piece.

Speaker:

If YouTube was four times as

incremental as Google said,

Speaker:

but your YouTube was crazy expensive,

Speaker:

it still might not be worth it

even though it's four times.

Speaker:

More.

Speaker:

Incremental than the platform was making.

Speaker:

And that's how I think a lot

about this with connected tv where

Speaker:

connected TV can be super powerful

and maybe more so than linear tv,

Speaker:

but if you can buy scatter

linear TV for a 10th

Speaker:

of the cost of CTV,

Speaker:

well it just has to be more

than a 10th as effective and

Speaker:

it's accreted, it's a positive.

Speaker:

So it becomes more of comparison

of a cost per than just a

Speaker:

blanket.

Speaker:

How incremental is something which I

always think is important to focus on and

Speaker:

call out.

Speaker:

To. Yeah, it's so good.

Speaker:

I mean measuring something in terms of

percentages can provide insights and help

Speaker:

make decisions, but ultimately

it's the cost per right.

Speaker:

Translate that into real dollars

to see if it makes sense.

Speaker:

100% agree with you,

Speaker:

but I think this also goes back

to and use your linear TV example,

Speaker:

and I still love TV and

connected TV and stuff. Again,

Speaker:

I'll use YouTube just because

I've got the numbers in my brain,

Speaker:

but with YouTube sometimes

we'll see a $5 CPM or a

Speaker:

$7 CPM in certain audiences

compared to other channels that are

Speaker:

15, 20, 30, 50, whatever.

Totally. And I'm like, well,

Speaker:

if we're reaching the right person

and if the message and offer are

Speaker:

good, how could this not work? And it's

one of those things where it's like,

Speaker:

okay, we're either one of those is

off, we're talking to the wrong person,

Speaker:

that's the wrong message,

Speaker:

or we're just not measuring it properly

and that's where we need to look at it.

Speaker:

So did you have a thought on that?

Speaker:

You another question on

MM here in just a second.

Speaker:

Yeah, yeah, totally. But it

made me think of the idea of,

Speaker:

I think the reason I'm starting to become

way more bullish on any channel that's

Speaker:

historically been hard to measure

where I think there's that arbitrage

Speaker:

opportunity of costs are still relatively

low because people haven't all moved

Speaker:

in because it's easy to attribute.

Speaker:

It'll be really interesting

with a house example,

Speaker:

does that inspire a lot

more YouTube buyers?

Speaker:

That's something that Google

should have put out way long ago,

Speaker:

but I think it would undermine

undermine search and that's their bigger

Speaker:

business. And I could do a whole

kind of rant and I'll save you that,

Speaker:

but the idea of incrementality first

measurement probably wouldn't be great for

Speaker:

the search business. So probably exactly,

Speaker:

haven't been able to make such a

good point that case on YouTube.

Speaker:

But you think about all the channels

that have historically been harder to

Speaker:

attribute,

Speaker:

that's where costs are deflated just

from a supply and demand perspective.

Speaker:

So when you can move in and get CPMs at

five to $7 and it's really effective,

Speaker:

but most people that are measuring

through attribution don't know it's really

Speaker:

effective, that's a huge win for certain

period of time until everybody's flood,

Speaker:

everybody and the costs go.

Speaker:

Up the market.

Speaker:

I'm sure there's a lot of people that

were not excited to see that study from

Speaker:

house like dang it, that means my costs

are going up. I don't like that at all.

Speaker:

So really good man.

Speaker:

So we talked about incrementality testing

and I think you can use tools like

Speaker:

House and then there are others.

Speaker:

We're just talking about work magic and

there's a number of others you can lean

Speaker:

into. Full disclosure,

they're pretty expensive,

Speaker:

but you can also do stuff on your own too.

Speaker:

If you've got someone that

can measure this stuff,

Speaker:

you can do a little bit of it on your

own. What about the MMM side of things?

Speaker:

What's kind of the easy way to start

there? Is there an easy way to start?

Speaker:

What do you recommend to people.

Speaker:

There? I don't know. I dunno if

there's an easy way to do anything.

Speaker:

I think, well, I guess

that's not totally true.

Speaker:

I think there's some ways to

run relatively easy incre tests.

Speaker:

So I think that's the

easier place to start.

Speaker:

Certainly you can always

ratchet up the scientific rigor.

Speaker:

I think the problem with looking

for an easy MM solution is

Speaker:

anybody could run a model with Robin or

there's a lot of open source packages,

Speaker:

but just because you can run a model,

Speaker:

it could say anything.

Speaker:

It's not necessarily rooted in this

can all of a sudden predict the future

Speaker:

and tell you exactly the

contribution from media.

Speaker:

Whereas incrementality can do

that a little more out of the box.

Speaker:

You may have wildly wide

confidence intervals,

Speaker:

but it answers the question.

It gives you the comparison.

Speaker:

I didn't do it in this market,

Speaker:

I did it in this market.

What is the Delta Media mix modeling?

Speaker:

You could build a model

to tell sort of any story.

Speaker:

The proof is sort of in the pudding of

if I do the thing that the model says,

Speaker:

does it change my top line?

Speaker:

Can I see over time that

when I listen to the model

Speaker:

that improves my top line?

Speaker:

So it's a lot easier to get started

with incrementality testing.

Speaker:

You can run poor man's match

market tests as I sort you can just

Speaker:

sort of pick,

Speaker:

some markets historically behave

similarly and there's certainly some risk

Speaker:

there, but with a model you might

think that it's an amazing model.

Speaker:

I just don't feel like there's a great

place to DIY that together without some

Speaker:

real scientific or statistical

rigor. Or if you do,

Speaker:

you've just got to try to prove it over

and over by taking some big swings. And

Speaker:

that's really,

Speaker:

I sort of feel like you can get away

with the kind of feel it sort of tests

Speaker:

without really running a true

incrementality test or model.

Speaker:

If you're a small enough business and

you spend a decent amount on Facebook,

Speaker:

maybe you're not willing

to turn off Facebook,

Speaker:

but are you willing to drastically

increase spend and see if you can feel

Speaker:

something at the top line? Okay, then

what happens if you cut it in half?

Speaker:

What happens?

Speaker:

And start to understand those curves on

your own is probably a less risky way

Speaker:

than trying to, I've never done

anything in R and I'm going to run

Speaker:

or done any sort of medium amount.

I'm going to try to run one.

Speaker:

That's probably a risky proposition.

Speaker:

Yeah, it's a really good insight. I'm

glad you answered the question that way.

Speaker:

I think, yeah,

Speaker:

leaning into the poor man's incrementality

test or just leaning really heavily

Speaker:

into a channel and measuring your top

line if you've got a small enough business

Speaker:

to look at that, but probably if

you're going to lean into MM M1,

Speaker:

you need a couple years of data and so

to be able to make some correlations and

Speaker:

you probably need to lean in to

someone or a tool with quite a bit of

Speaker:

experience because you can do that astray.

Speaker:

And on your comment on cost too.

Speaker:

I mean it's all relative and a lot of

times where you're going to need a medium

Speaker:

mix modeling is when you're spending

a significant amount in a significant

Speaker:

number of channels,

Speaker:

which you're probably only doing

if you are spending a lot total,

Speaker:

which you're probably only doing if your

revenue can support that high level of

Speaker:

spend,

Speaker:

which means that a tool may not be

all that expensive relative to the

Speaker:

opportunity you could derive from

it, which is where I always net out.

Speaker:

So I'm paying 10 or 20

grand for a tool monthly,

Speaker:

but it's allowing me to

redeploy millions in ad spend.

Speaker:

And it totally in completely

makes sense. So Tom,

Speaker:

this has been fantastic.

I'm just watching the clock.

Speaker:

I know we're kind of coming

up against it, but one,

Speaker:

I recommend people follow you on LinkedIn.

You put out some awesome content.

Speaker:

I love reading it.

Speaker:

Thank.

Speaker:

You. People should definitely follow

you on LinkedIn and you are, is it Tom,

Speaker:

what is your handle on LinkedIn?

You are Thomas B. Leonard.

Speaker:

Thomas B. Leonard. That's

probably confusing.

Speaker:

I'm very self-conscious of LinkedIn, so

I'm glad to thank you for saying that.

Speaker:

I think it's good, man. I think it's

really good. I like it a lot. Yeah.

Speaker:

Yeah, it's been fun to start

doing connecting with folks.

Speaker:

Definitely an area that had a lot

of excitement and passion for,

Speaker:

it's fun to have these

sort of conversations,

Speaker:

so I appreciate you reaching out a

while ago and that we could connect.

Speaker:

Absolutely.

Speaker:

Man. Absolutely. So then if

other people were like, Hey,

Speaker:

I just want to talk to Tom because maybe

you can help my brand or my business,

Speaker:

how can they connect with you and who are

you looking to or who do you feel like

Speaker:

you can help?

Speaker:

Yeah, definitely appreciate that.

Yeah, reach out on LinkedIn.

Speaker:

I spend time there. I love reading

everybody's thoughts and content. So yeah,

Speaker:

reach out on LinkedIn mostly we work

with consumer facing brands that

Speaker:

are trying to understand where to

put the next dollar or where to pull

Speaker:

in the scenarios. They have to really

kind of rescue people from attribution,

Speaker:

trying to better understand where they

can get more with their ad dollars.

Speaker:

I think to your point that you teed

up now is such an interesting time or

Speaker:

anytime that there's margin pressure,

Speaker:

there's more scrutiny

on a marketing budget.

Speaker:

Really want to try to help

empower marketing teams to

feel more confident with

Speaker:

what they're doing and ultimately the

finance teams to feel more confident with

Speaker:

what marketing team is doing. Hundred

percent. That's where I love to plug in,

Speaker:

but also just love to talk about this

stuff probably more than I should.

Speaker:

So always open to the conversation.

Speaker:

Yeah, I talk about that a lot.

Speaker:

I've read analytics and measurement

books on vacation and my wife

Speaker:

is like, what is wrong with you? And I'm

like, it's interesting. I don't know.

Speaker:

I like it. And so totally, we are

just a different breed I suppose,

Speaker:

but I love that.

Speaker:

And then I think this is a great way to

end it where if I've got an extra dollar

Speaker:

to spend on marketing, where do I put

it? If I need to cut a dollar of spend,

Speaker:

where do I cut it from?

Speaker:

And that's really what

this approach is about MMM

Speaker:

and incrementality. And so

I think their necessities,

Speaker:

I think attribution is broken and or

misleading in so many different ways.

Speaker:

There's some correlations there, so we

don't have to throw it out completely,

Speaker:

but I do believe you need to lean

into MMM and incrementality for short.

Speaker:

So connect with Tom on LinkedIn.

And with that, we'll wrap.

Speaker:

Tom's been fantastic. Thanks for the

time, the insights and the energy. Yeah.

Speaker:

Thanks so much Brett

time. Glad to connect.

Speaker:

Absolutely. And as always, thank you for

tuning in. We'd love to hear from you.

Speaker:

If you found this episode helpful,

Speaker:

someone else in the D two C space or

marketing space, and you think, man,

Speaker:

they got to listen to this, please

share it. We mean the world to me.

Speaker:

And with that, until next

time, thank you for listening.

Links

Chapters

Video

More from YouTube