In this insightful episode of the E-commerce Evolution Podcast, host Brett Curry sits down with Tom Leonard (https://www.linkedin.com/in/thomasbleonard), a fractional marketing leader who specializes in operationalizing Media Mix Modeling and incrementality testing. They dive deep into the often confusing world of marketing measurement. Tom and Brett will debunk myths about attribution and we reveal what truly drives customer acquisition.
For ecommerce brands struggling to understand where their marketing dollars are actually working, this conversation offers practical insights on how to move beyond misleading platform metrics.
—
Sponsored by OMG Commerce - go to (https://www.omgcommerce.com/contact) and request your FREE strategy session today!
—
Chapters:
(00:00) Introducing Tom & Marketing Measurement
(06:30) Understanding Multi-Touch Attribution (MTA)
(12:22) The Case for Incrementality Testing
(22:20) Exploring Media Mix Modeling (MMM)
(27:30) Navigating Budget Cuts and Marketing Spend
(32:17) Understanding Incrementality Vs. Attribution
(35:45) The Importance of Cost Per Incremental
(40:16) How to Get Started with MMM
(44:09) Final Thoughts
—
Connect With Brett:
Relevant Links:
__
Past guests on eCommerce Evolution include Ezra Firestone, Steve Chou, Drew Sanocki, Jacques Spitzer, Jeremy Horowitz, Ryan Moran, Sean Frank, Andrew Youderian, Ryan McKenzie, Joseph Wilkins, Cody Wittick, Miki Agrawal, Justin Brooke, Nish Samantray, Kurt Elster, John Parkes, Chris Mercer, Rabah Rahil, Bear Handlon, Trevor Crump, Frederick Vallaeys, Preston Rutherford, Anthony Mink, Bill D’Allessandro, Bryan Porter and more
How much is media contributing relative
to customer base is a really nice place
Speaker:to start.
Speaker:And the benefit of running
incrementality and media mix modeling is
Speaker:informing the model with
some of that causal data.
Speaker:Well, hello and welcome to another edition
of the E-Commerce Evolution podcast.
Speaker:I'm your host, Brett
Curry, CEO of OMG Commerce.
Speaker:And today we have got
a doozy of an episode.
Speaker:We're talking about the three
horsemen of measuring your
Speaker:marketing effectiveness. We're
talking MTAs Multitouch attribution.
Speaker:We're talking M'S. Media mixed
modeling. We're talking incrementality.
Speaker:It's going to be nerdy,
Speaker:but I also promise you it's going to
be practical and it will make you more
Speaker:money. And so we'll hopefully
make it fun as well.
Speaker:And so my guest today is Tom Leonard.
Speaker:We are LinkedIn friends first.
Speaker:So I saw Tom on LinkedIn posting about
incrementality, talking about MMM,
Speaker:throwing shade on certain tools and stuff
like that on LinkedIn. And I'm like,
Speaker:this is my type of guy. So I reached
out, we had a call, and then we're like,
Speaker:Hey, we got to record a podcast.
Speaker:Let's create some insights
for people on the pod.
Speaker:And so Tom is a fractional
marketing leader.
Speaker:He's operationalizing MMM
and incrementality testing,
Speaker:and I'm delighted that he's my guest
today. So Tom, with that intro,
Speaker:how's it going? And welcome to the show.
Speaker:Good. Thanks for having me, Brent.
Excited to be here. And yeah,
Speaker:some of my favorite things to talk
through, so excited to do it. Good stuff.
Speaker:It's good stuff, man. So briefly,
Speaker:before we dive into the
meat of the content here,
Speaker:what's your background and
how did you become a guy who's
Speaker:operationalizing MMS and incrementality?
Speaker:Yeah. And what does that even mean?
Speaker:That's a good point.
Speaker:For sure. Yeah, totally. Yeah.
Speaker:So spent most of my career thus far on
the agency side at performance agencies.
Speaker:And I'd say the crux of
how I got to where I'm now,
Speaker:or I've been reflecting back a little
bit more on the why I have such a passion
Speaker:for measurement. And I was at
a pretty hardcore DR agency,
Speaker:and it was right shortly after TRUBY
for Action came out when YouTube was
Speaker:starting to invest in, DR.
Speaker:Moved into a new role we had created
with a centralized group of basically
Speaker:people who had different areas of subject
matter expertise and a few analysts
Speaker:that ran tests across a
pretty large client base.
Speaker:And I was our YouTube SME,
Speaker:and worked with a couple
analysts to run a bunch of tests.
Speaker:And really it was to evangelize how to,
Speaker:and is YouTube a platform to drive growth?
And it was really interesting
Speaker:because I started spending a lot of time
on YouTube and then also connect to TV
Speaker:and broader programmatic video.
And it was this really interesting,
Speaker:for me, the biggest learning was less
about how to make YouTube as effective as
Speaker:possible,
Speaker:but more how to help brands think about
demand creation as opposed to just
Speaker:demand capture. And frankly,
Speaker:the difficulty of getting brands
to leverage YouTube relative
Speaker:to connected tv,
Speaker:because YouTube sat so close to Google
ads and therefore last click attribution
Speaker:and see tv, you couldn't click
and was sexier in a deck.
Speaker:And it was just this sort
of recognition of the
Speaker:irrational kind of human behavior just
in any sort of industry or any thing
Speaker:in life.
Speaker:But it sort of helped frame up this
idea of you really have to do more than
Speaker:just, I don't know,
Speaker:represent logic or rational arguments.
You really have to also
Speaker:bring the easy to understand
clear data. And that's,
Speaker:I think what draws me to incrementality
testing specifically and why
Speaker:that's sort of the backbone
of a lot of what I do now.
Speaker:And I think I use the word
operationalizing, NMM and
incrementality testing.
Speaker:And really what I mean by that is a lot
of people will run medium mix models or
Speaker:run incrementality tests,
Speaker:but oftentimes they'll sit in a slide
or in a report to be shown once,
Speaker:but never to be looked at again.
Speaker:And so what I'm really trying to do
with brands now is how do you build a
Speaker:framework and a repeatable methodology
to get insights from tests,
Speaker:but not just leave them as
insights but to take action?
Speaker:Because the only way that you create
value from any of these sort of testing
Speaker:methodologies and measurement
methodologies is by
acting on the insights.
Speaker:And so that's sort of what I mean by my
funky little headline of those words.
Speaker:Yeah, it's so good, man.
Speaker:And it's one of those things where data
really doesn't matter if you don't take
Speaker:the right actions from it.
And what's so interesting,
Speaker:and our paths are similar in that
I got my start in actually TV and
Speaker:radio and doing traditional media, and
then I got into SEO and paid search,
Speaker:but I loved video. Video was my
thing, but I love paid search as well.
Speaker:And then when TrueView and TrueView
for Action came out, I was like, whoa,
Speaker:these are all my world's colliding.
Speaker:This is.
Speaker:Video and there's some search components,
Speaker:at least some search intent involved
there. And it's direct response.
Speaker:I've always been a direct response guy.
Speaker:I believe that marketing
should drive an outcome, right?
Speaker:Advertising should drive
a measurable outcome,
Speaker:and that should be measured in terms
of new customers and profitable new
Speaker:customer acquisition. And
what's really interesting, Tom,
Speaker:and I think this kind of feeds into
the conversation we're having today.
Speaker:There was a period of time, so I
grew up reading some of the classics.
Speaker:So David Ogilvy of course, but John
Cap's tested advertising methods,
Speaker:Claude Hopkins Scientific Advertising.
Speaker:And they would do things like they would
run and add in a newspaper or magazine
Speaker:and people would clip a
coupon and bring it in,
Speaker:or they would call a certain number and
they would track it and they would have
Speaker:codes and stuff.
Speaker:And I remember thinking once I got
into e-commerce, I was like, oh man,
Speaker:we've got so many tools. The world is
so clear now we have every piece of
Speaker:data at our disposal.
Speaker:And now the more I've gotten into it
and the more I've matured, I'm like,
Speaker:we've got more data. But I don't
know that we've got more insights,
Speaker:and I don't know that we've
got any more clarity. In fact,
Speaker:there's maybe more confusion.
Speaker:And I think it goes back to
what you said a minute ago,
Speaker:this idea of demand generation
versus demand capture.
Speaker:We're really good at measuring channels
and campaigns that are demand capture,
Speaker:meaning they're capturing
demand that's already out there.
Speaker:That's harder to measure
the demand generation,
Speaker:which is usually where the magic happens.
Speaker:And so super excited to dive in here.
Speaker:I think what might be useful
is let's talk about what
Speaker:are these kind of three horsemen that
I laid out there, MTAs, multitouch,
Speaker:attribution, and incrementality.
So let's start with MTAs first.
Speaker:So Multitouch attribution tools,
Speaker:what are they and what
is your take on them?
Speaker:Yeah, big question. Great
question. Yeah, I mean,
Speaker:MTA been around for a while,
Speaker:different flavors and ways
of trying to make it work,
Speaker:especially as so much has changed
in privacy and the tech and tracking
Speaker:landscape.
Speaker:But ultimately the goal is to try
to give fractional credit to all the
Speaker:touchpoints along a customer journey with
a recognition that the last touchpoint
Speaker:click or last impression is
ultimately not what drove that person
Speaker:to purchase.
Speaker:That may be the last or the only thing
that you might see in something like
Speaker:Google Analytics or your analytics suite.
Speaker:But there's this general recognition
that that is not what drove the purchase.
Speaker:So MTA, the kind of promise, which I
ultimately think is a failed promise,
Speaker:is whether all the different touch
touchpoint and then how can you
Speaker:value those differently. So
maybe you use first touch,
Speaker:maybe you use even distribution. The
idea of data-driven attribution was the
Speaker:holy rail or the Promise many years ago,
Speaker:and I guess still to a
degree for some is like,
Speaker:how do you know this channel was more
additive or more necessary and therefore
Speaker:should get more credit than that channel?
Speaker:Which I think makes a
ton of sense in promise.
Speaker:I think in reality it's really hard
and I would argue impossible to do,
Speaker:especially as a lot of the ability to
track users at a one-to-one level degrades
Speaker:generally my perspective,
I'm very bearish on MTA,
Speaker:so that'll probably come
through pretty strongly.
Speaker:But I guess I don't think the toothpaste
is going back in the tube in terms of
Speaker:the ability to track a customer across
all these different touchpoints,
Speaker:especially as the ability to
track through or impression based
Speaker:touchpoint erodes. And then you
really get reliant on clicks,
Speaker:which I think then leads to a lot of
all the issues that just last click in
Speaker:general has.
Speaker:So I think it's really hard to
make a compelling case for MTA.
Speaker:I've seen too many brands,
Speaker:especially trying to
build MTA tools internally
Speaker:and just be a huge time and resource
suck. And then when you ask to compare,
Speaker:show the multi-touch view versus
last click, it's like, I don't know,
Speaker:80 or 90% only had one touch
point anyways, that's all
that MTA model could see.
Speaker:So is it really that much
more useful than last click?
Speaker:It's sort of multi-touch when that can
be measured, but usually it can't be.
Speaker:Yeah, and It never really answers
the causality question either,
Speaker:which we'll get to when we
talk about incrementality.
Speaker:And I always kind of tell this,
Speaker:I think the short story of why MT A
isn't really viable anymore as all the
Speaker:tracking and privacy changes.
Speaker:But I think the slightly longer story
is the kind of recognition that just
Speaker:because an ad was shown or a
click occurred doesn't mean that
Speaker:that medium was needed or
that channel was needed.
Speaker:It doesn't answer the causal question,
Speaker:what would've happened
without this ad running?
Speaker:Did somebody just happen to use multiple
touchpoints as navigation or was it
Speaker:more convenient to click on one of
these ads that happened to be served?
Speaker:But if you're not comparing that to some
sort of control group to really hard
Speaker:to assign causality to the fact
that there just was a touchpoint.
Speaker:Yeah, it is so good. And it's one of
those things where I remember again,
Speaker:early on,
Speaker:you would look inside of Google ads or
you look inside of Meta or was back when
Speaker:it was Facebook only, and you
were like, the data's here.
Speaker:I see row ads and I see clicks and
I see performance and all that.
Speaker:Then you realize, well, wait a
minute, this isn't fully accurate.
Speaker:If I add the two together,
that's double my total revenue,
Speaker:so I can't just rely on
what's in the platform.
Speaker:And that got worse as I was 14 was
introduced and other privacy changes were
Speaker:made. But then MTA came
along and it's like, oh,
Speaker:finally we're going to get to see the
full picture. It's going to decipher,
Speaker:decode the shopping journey,
Speaker:and we're going to finally see with a
keen eye in perfection exactly what caused
Speaker:this ad or what caused this purchase
to happen. And then we finally realized
Speaker:MTA is maybe just a third
option. It's like, okay,
Speaker:Google's imperfect, Meta's
data's imperfect, and then mt A,
Speaker:it's just imperfect too.
Speaker:So now we just got three imperfect
things to look at and make
Speaker:decisions from.
Speaker:And in some ways it leads to more
confusion than it leads to clarity.
Speaker:And now I don't want to wholesale discard
Speaker:MTAs because I do believe there's some
helpful insights that can be gained
Speaker:there,
Speaker:but it's incomplete
and incomplete at best.
Speaker:And one of the best analogies I've heard,
and this actually comes from Ben Ter,
Speaker:who's also a LinkedIn friend,
but I met him in person as well,
Speaker:but he talks about this analogy of, Hey,
Speaker:if we're trying to measure what
caused people to watch this
Speaker:movie at our movie theater,
Speaker:and we look at all these
results and 30% say they saw a
Speaker:billboard for our movies,
20% say they saw a TV ad,
Speaker:but you know what? A hundred percent
say they saw the poster on the
Speaker:door. So we're like,
let's just cut everything.
Speaker:Let's just do the poster at the door
and that's it. And you're like, well,
Speaker:wait a minute. Everybody saw it.
Everybody was walking in the door.
Speaker:But the movie poster is not
what caused someone to purchase.
Speaker:It was the billboard and the TV
and some of the other things,
Speaker:word of mouth and other things
that caused them to come in.
Speaker:And so this idea of causality,
super, super valuable.
Speaker:So that really leads us to incrementality.
So talk about incrementality.
Speaker:What is it and why are you on
a quest to operationalize it?
Speaker:Yeah, it's really the best way,
Speaker:if not the only way to
establish that a causal
Speaker:portion that we've been talking about.
It has a distinct control group,
Speaker:so it has a counterfactual,
Speaker:it has what would've happened
without this intervention,
Speaker:whatever that intervention is.
Speaker:And there's a handful of ways to derive
that counterfactual that control.
Speaker:The most common would be geographic
based. So like a match market test.
Speaker:I've got this market over here that
historically has behaved similarly to this
Speaker:market over here. I can
see that in an AA test,
Speaker:the lines sort of move similar
to one another. They're not,
Speaker:if they're influenced by outside
factors, they're influenced.
Speaker:In what's an AA test for
those who don't know.
Speaker:Before an intervention happens.
Speaker:So just over time are those lines
essentially moving together?
Speaker:Are external factors or stimuli equally
impacting both sides of that test
Speaker:so that you can feel confident that
when you do intervene and it becomes
Speaker:comparing A to B,
Speaker:the delta is what was a
result of that intervention.
Speaker:So oftentimes it's my Atlanta
Speaker:and I don't know Memphis,
Speaker:maybe some other midsize city that
you've done this market matching for.
Speaker:Historically, they both
look like this on a line,
Speaker:all of a sudden you turn off
ads on Facebook in Atlanta,
Speaker:what happens to your top line that
Delta is what was attributed or
Speaker:should be attributed to
advertising in Atlanta.
Speaker:Whereas the flip side of that would be
attribution would say basically anything
Speaker:that was attributed to that could
be attributed to that would really,
Speaker:it should just be the gap between a
world where that ad does not exist
Speaker:compared to a world where that ad
does exist. We can't take credit for
Speaker:everything.
Speaker:We can only take credit for as much
above and beyond what would've happened
Speaker:anyways. And so that's the
basis of incrementality testing.
Speaker:There's other ways to do it.
Speaker:If you use a Facebook or Google
conversion lift study because they own
Speaker:that auction or anybody
that owns an auction,
Speaker:they can do that hold out
for you at a user level.
Speaker:They can track all of those users
regardless of if you serve an ad.
Speaker:Good examples are maybe easier to
describe in a first party data capacity.
Speaker:If you're running email, you may blast
all of your customers and say, Hey,
Speaker:I sent an email to all my
customers and this many purchased.
Speaker:They went back to the website or
clicked it. But if you just said, Hey,
Speaker:I'm going to serve just to odd
number of customer IDs and not to
Speaker:even number customer IDs,
I can then just compare,
Speaker:forget about who clicked on ads,
Speaker:who did anything.
I'm just going to look at my backend.
Speaker:I know I exposed these users,
but not these users 50 50 split.
Speaker:They've historically kind
of done the same thing.
Speaker:All I did was even an odd and just
measuring the difference between those two
Speaker:groups.
Speaker:So really any way that you can
establish a true control that
Speaker:passes that AA test. So
before you intervene, do they
continue to look similar?
Speaker:Are they influenced at the same rate so
that you can feel confident that when
Speaker:you do intervene with new
media, retracting media,
Speaker:some new sort of test that you are
confidently comparing to what would've
Speaker:happened in a world
without that intervention?
Speaker:Yeah, yeah.
Speaker:It's applying the scientific
method with some rigor behind
Speaker:what happens when I turn this channel on,
Speaker:or what happens when I
turn this channel off?
Speaker:What is the actual impact of this channel?
Speaker:And what's interesting is I
remember back in my early days
Speaker:of being in the advertising world,
Speaker:this was when online stuff was
just getting kind of warmed up.
Speaker:I was talking to this furniture store
owner and I'm like, Hey, what do you do?
Speaker:Do you invest in radio ads?
Do tv, do you do newspaper?
Speaker:And so as I went through Themm like,
Hey, do you do radio ads? And he is like,
Speaker:yeah, I mean, yeah, I sort of do.
And I'm like, newspaper's like, yeah,
Speaker:there's a big sale, something will
happen. I'm like, well, what about tv?
Speaker:And he said, yes. And his
eyes lit up and he is like,
Speaker:when I run TV ads, I feel
it. People walk in the door,
Speaker:it happens. And I remember early on
in my online career thinking, man,
Speaker:that was so unsophisticated. Did
that guy really know what's going on?
Speaker:But now looking back, I'm like,
yeah, that's maybe all that matters.
Speaker:That is incrementality in a real loose
easy just to observe with your eyes think
Speaker:because you had one. Totally.
Speaker:Which I think people
take for granted. Yeah.
Speaker:They do.
Speaker:Yeah.
Speaker:That's not exciting. That's not
like, where's all your data?
Speaker:It's in my cash register.
That's where all the data.
Speaker:Is, especially for smaller brands,
Speaker:when you have the ability
to feel if something's
Speaker:working or not working,
Speaker:if you double spend in something that
you think is working really well because
Speaker:attribution says it's working really well,
Speaker:and all of a sudden
your cash just doubles,
Speaker:even though your attributed number
scales linearly, something has to give,
Speaker:right?
Speaker:And what has to give is it wasn't really
causing any additional top line growth.
Speaker:It was just really good at
getting the attributed credit.
Speaker:So I think the feeling
it in the p and l is
Speaker:definitely overlooked.
Speaker:It's valid, and it is overlooked
though. You're a hundred percent,
Speaker:especially now that we have
so many tools at our disposal.
Speaker:And I think another way to look at
this, and look, I'm a Google guy,
Speaker:YouTube and Google is kind of where
I really got my start in online.
Speaker:Marketing.
Speaker:But listen, branded search is a
perfect example here. What happens,
Speaker:we see this all the time.
Speaker:What happens if you turn branded
search completely off? Now, I believe,
Speaker:and this is top of front of the podcast,
Speaker:there are strategic ways to use branded
search and there's ways to run it and
Speaker:not waste money, but a lot of people
could shut it off and nothing happens,
Speaker:nothing. Maybe sales get in a little bit,
Speaker:but you take meta meta's really working
and you shut it off and you feel it.
Speaker:Sales go down and that's
an incrementality.
Speaker:Same is true for YouTube if you're doing
YouTube the right way. And so yeah,
Speaker:I really like this. And one
kind of anecdote here to share,
Speaker:we just did a test with Arctic,
Arctic coolers, Yeti competitor,
Speaker:my favorite cooler, my favorite drinkware
as well. And so they wanted to see,
Speaker:Hey, can YouTube drive an incremental
lift at Walmart? So they had just
Speaker:gotten into most Walmart
stores, coast to coast.
Speaker:So we did exactly what you laid out
there. We had a 19 test markets,
Speaker:19 matched control markets.
So similar markets.
Speaker:So think like a Denver and a
Kansas City or the example,
Speaker:use Atlanta and whatever else
that's kind of comparable. And hey,
Speaker:let's run YouTube in one
and not in the other.
Speaker:And let's measure then the
growth in Walmart sales,
Speaker:and let's do a comparison
between the two in Walmart sales.
Speaker:And it was remarkable. It
was about an eight week test.
Speaker:We had three test regions, so 19
markets, but three test regions,
Speaker:test region. One, we saw an average
of 12% lift in Walmart sales.
Speaker:The test region two was like 15% lift.
Speaker:And then our final test
region was 25% lift.
Speaker:And there were some standouts,
Speaker:like Oklahoma City was up 40% and Salt
Lake City was up 48%. But it was one of
Speaker:those things where, okay, now we
look at that and we can say, okay,
Speaker:YouTube had a big impact. And
what's also interesting, Tom,
Speaker:is we just ran the YouTube portion at OMG.
Speaker:They also did a connected TV test
in other markets, not related,
Speaker:didn't see a lift, didn't
see a measurable lift.
Speaker:And so it could be lots of
that was not to throw shade on
Speaker:CTVI like CTV,
Speaker:so maybe they just did a wrong or
wrong creatives or who knows what.
Speaker:But it's one of those things
where it's like, okay,
Speaker:if you do this the right way,
you should see an impact.
Speaker:And I think touching on the
piece that I didn't mention,
Speaker:the other beauty or value of
incrementality testing relative to
Speaker:attribution or mt a is the ability
to see beyond your.com to be able to
Speaker:see what's happening on third parties
like Amazon, what's happening in store.
Speaker:If you get that data own an operated
store or if you can get that through
Speaker:wholesale data, it really simplifies.
Speaker:There's so much complexity.
And I think that's, again,
Speaker:one of the rubs that I have
with MTA is all of them,
Speaker:all of the data you have to
wrangle together to try to
Speaker:patchwork this kind of story together.
Speaker:Whereas in incrementality testing,
it's pretty straightforward.
Speaker:It's what did I spend and how
did I run that spend in these by
Speaker:market by day or by week, and what
was my sales? What were my sales?
Speaker:What were my new customers or whatever
metric I'd want to look at with that same
Speaker:granularity and same dimension.
Speaker:And that's really it because you're
really just trying to understand the
Speaker:relationship that calls the
relationship between spend and outcomes,
Speaker:all that kind of muddy middle
in the middle, trying to
get it at the user level,
Speaker:which again, not going back into
the tube really simplifies things.
Speaker:Yeah, it does.
Speaker:And another thing that was
kind of interesting that
came a light doing this test
Speaker:for Arctic is all of the ads we
tagged with available at Walmart,
Speaker:shop at Walmart, find on the
shelves and Walmart, whatever.
Speaker:We measured everything
though in those markets.
Speaker:So you could look at Walmart sales,
online sales, so the.com and Amazon.
Speaker:And what's interesting is the
push to Walmart really worked.
Speaker:It's a reminder of what you ask someone
to do in an ad is what they're going to
Speaker:lean towards. Because
in some of the markets,
Speaker:we didn't see that much of an online lift.
Speaker:We saw some clicks and stuff like
that, but the Lyft was at Walmart.
Speaker:But we also saw a pretty
strong lift at Amazon as well,
Speaker:because I think that just speaks to,
Speaker:there's some people that are just going
to buy everything from Amazon right
Speaker:there, tell 'em to go online value pro
proposition. Is it on Amazon? Yeah, yeah.
Speaker:Yeah. Here in a day or two, it's hard.
Speaker:To beat, dude. It's hard to beat
same price in a couple days.
Speaker:I don't have to leave my house. But
yeah, really, really interesting.
Speaker:And so we'll circle
back to that of course,
Speaker:but let's talk about then
MMM or media mix modeling.
Speaker:What is that? How are you using that?
Speaker:And then how does that kind of relate to
incrementality testing? Because again,
Speaker:going back to your tagline, Tom, you
did not say operationalizing NTAs.
Speaker:You said operationalizing m
and ms and incrementality.
Speaker:So what is MM and how does
that pair with incrementality?
Speaker:Yeah,
Speaker:basically a big correlation exercise
trying to suss out without a true kind of
Speaker:holdout group,
Speaker:what is the impact and contribution of
each media channel and also what would
Speaker:happen without media.
Speaker:So trying to suss out a lot of the
same questions as incrementality,
Speaker:but basically using correlation as
opposed to having a true holdout group.
Speaker:So basically,
Speaker:and I'm sure all the hardcore MMM people
and data scientists will thumbs down
Speaker:this or whatever you can do to podcast,
but hey, in this period of time,
Speaker:sales went up and nothing could really
explain that other than the fact that
Speaker:TikTok spend went up and essentially
doing that at a mass scale over longer
Speaker:periods of time trying to take into
account anything that could explain that.
Speaker:So you'll always kind of flag it with
these are promotions that happen,
Speaker:it should because you're going to give
a model at least like two years worth of
Speaker:data or two years worth of data,
Speaker:it'll bring in seasonality and try to
understand those sort of trends. So it's
Speaker:trying to pull out if not
seasonality, if not promotions,
Speaker:if not some other things
that we are flagging.
Speaker:And it wasn't price reductions,
it wasn't all these pieces,
Speaker:what was happening in media
that could explain that change.
Speaker:And so that's ultimately
what MMM is doing.
Speaker:It's a big correlation exercise,
Speaker:figuring out roughly what is the channel
contribution to a top line revenue or
Speaker:order number and what's really important.
Speaker:I think the nicest part or the best
first step with M is trying to get an
Speaker:understanding of a base,
Speaker:which is what it's going to be called or
intercept what without the presence of
Speaker:ads,
Speaker:does this model think that my sales would
be such that I can then calculate not
Speaker:a total CAC of just looking at
total new customers divided by cost,
Speaker:but incremental to media
or remove base from
Speaker:that equation,
Speaker:how many conversions were contributed
because of media as this model sees,
Speaker:which no model is going to be perfect,
Speaker:no measurement method
is going to be perfect,
Speaker:but it's a really nice
place to start to say,
Speaker:I knew I couldn't account all
new customers to advertising,
Speaker:but what's a good number to use or
to start with? Well, it looks like,
Speaker:and this will depend on the maturity of
the brand, but a really mature brand,
Speaker:I mean super mature brand,
Speaker:the big CPGs might be like 99% base
smaller brand might be something
Speaker:like 50% because you've got
this word of mouth flywheel,
Speaker:you've got product market fit,
Speaker:but trying to get an understanding of how
much is media contributing relative to
Speaker:customer base is a really
nice place to start.
Speaker:And the benefit of running
incrementality and media mix modeling is
Speaker:informing the model with
some of that causal data.
Speaker:You see that a lot and there's a
really powerful feature of media mix
Speaker:modeling is saying, Hey, yes,
that's a correlation exercise,
Speaker:can't pull everything out,
Speaker:but let me inform the model or at least
restrict the priors it can use or the
Speaker:coefficient, whatever
you want to call 'em,
Speaker:what it's searching for to try to find
a fit in this model and say, well,
Speaker:I did a hold out test. I know
you don't have the causal data,
Speaker:but we ran this in this channel and that
channel and helping that restrict the
Speaker:model and giving it data that it can't
have without that human intervention can
Speaker:be a really powerful flywheel.
Speaker:So using your incrementality test data,
Speaker:feeding that back into your MMM
model to make it more accurate and
Speaker:more causal and make that correlation.
Speaker:Stronger.
Speaker:Because the two things that are really
like you're really trying to get,
Speaker:but you don't get with Multi-Tech
attribution or attribution in general.
Speaker:And you do get with the combination of
media mix modeling and incrementality
Speaker:testing is the incremental impact,
Speaker:the causal impact of what
would've happened without
the presence of ads as well
Speaker:as the diminishing returns curve,
Speaker:which we know can be really
powerful and important too,
Speaker:is what has happened over time as I
spend in that sort of a feature of big
Speaker:feature of media mix modeling
is understanding where
are you on a diminishing
Speaker:returns curve? Is there
if I keep spending more,
Speaker:I know it's not going to scale linearly,
Speaker:but are there channels
that diminish faster?
Speaker:Is there more headroom in other channels?
Speaker:And it really becomes this
true optimization game of
where do I put the next
Speaker:dollar? Ultimately the
question that every marketer,
Speaker:every finance team is
trying to answer is, Hey,
Speaker:if I find $20,000 into couch
cushions, where do I put it?
Speaker:And if I need to give back $20,000,
where do I pull from to have.
Speaker:I want to hang out at your house and
look at your couch cushions and find 20
Speaker:grand? That's.
Speaker:Great. Yeah, it's easy to
give it back, but yeah, right.
Speaker:We're trying to figure out what is going
to be the least impactful if I have to
Speaker:give the money back and cut budgets
and where is it going to be the most
Speaker:impactful if I have another $20,000?
Speaker:Because the answer is not going to be
found in what has the highest or the
Speaker:lowest ROAS in an attributed
view. And in fact,
Speaker:that can have the complete
opposite impact that you want.
Speaker:Yeah, yeah, it's really great.
Speaker:So I want to actually talk about
that point in a minute where
Speaker:if you've got cut budgets,
which hey, listen,
Speaker:there's been some uncertainty even as we
record this, tariffs up, tariffs down,
Speaker:markets up, market down, whatever
consumer sentiment is all over the place.
Speaker:So if things get a little bit
tight, what are we going to do?
Speaker:We can't slash marketing,
we can't slash growth.
Speaker:I think that sends you
into a death spiral,
Speaker:but we might have to get pull
back and get more efficient.
Speaker:And so let's talk about that
actually for a little bit.
Speaker:So where can you be led astray?
Speaker:I think you just made a post
on LinkedIn about this, right?
Speaker:Where you start looking at performance,
which feels like the smart thing to do,
Speaker:looking at ROAS and whatnot, and
you're like, well, great, well,
Speaker:let's just cut the lowest ROAS
campaigns and channels. We'll be fine.
Speaker:How does that lead you astray?
Speaker:And if you want to talk about your
specific example to help illustrate these
Speaker:points, that'd be great.
Speaker:Yeah, totally.
Speaker:I think the other one you're referring
to is I think branded search,
Speaker:which we were talking about
earlier. And I love using both a,
Speaker:because it can be really, if a brand
is spending a lot of money there,
Speaker:it can be a really great place to go
find those savings without impacting top
Speaker:line. But also frankly, it's
really easy to understand.
Speaker:I think most people understand that
up and down the organizational chart
Speaker:across departments, everybody sort
of understands the idea of, Hey,
Speaker:if somebody's already
searching for my brand,
Speaker:do I need to pay to get that
click and that conversion?
Speaker:And I found that just the fact that
it's easy to understand can be a
Speaker:really good gateway to incrementality
testing because it's easy to get buy-in.
Speaker:Everybody understands that idea,
Speaker:whereas it may be more challenging
to express that idea in
Speaker:other types of campaigns.
But branded search is a good example,
Speaker:and the example that you're referring to,
Speaker:kind of a midsize brand that I was
working with went through that exact
Speaker:exercise, had to cut budgets.
Speaker:They looked at up and down the campaigns
they were running. It was like, Hey,
Speaker:we just got to make the best decision
we can with the best available data.
Speaker:They were basically running p max
non-branded search and branded search and
Speaker:p max and branded search where had
the best attributed roas Best CPA
Speaker:non-brand was really hard to justify in
a lower budget kind of environment based
Speaker:off the attribution data cut that leaned
a little bit more into branded search
Speaker:as a percentage of their budget.
And over the next couple months,
Speaker:new customers in total revenue
was declining despite the
Speaker:attributed ROAS and CPA
looking even better than ever.
Speaker:And that's where was brought
in, looked at all these things,
Speaker:saw the loose correlation to
non-brand and new customer
Speaker:acquisition and top line,
Speaker:just the general skepticism that
many have around branded search,
Speaker:especially in a low
competition environment,
Speaker:which they were in. There weren't many
competitors in the auction that we
Speaker:could see in Auction Insights. So yeah,
Speaker:ran a very blunt instrument
match market test,
Speaker:which at a brand of that size and for a
branded search I don't think is ever a
Speaker:bad idea. And yeah, no
impact to branded search.
Speaker:It was about 20% of their budget,
Speaker:which was substantial that you
can either make the decision,
Speaker:I'm going to put that 20% back in
my pocket or save it for a rainy day
Speaker:or give it to some other
place in the org or say, Hey,
Speaker:I'm going to redistribute this to
something that I see in correlation
Speaker:data that might help
drive top line backup.
Speaker:Let's reinvest that in non-brand as
opposed to keeping it in branded. Again,
Speaker:complete opposite of what
attribution would say.
Speaker:And you see that a lot frankly with
branded search is an easy one to pick on.
Speaker:Same with retargeting,
Speaker:but really anything that's especially
challenging with the black box
Speaker:solutions that blend,
Speaker:and I'm sure we could do a whole talk
show on p max Advantage plus some of the
Speaker:things that bundled together historically
radically different levels of
Speaker:incrementality can be a real challenge
when you're then measuring on
Speaker:attribution. But yeah, a
ranty way of saying yes,
Speaker:finding areas to cut oftentimes
if you follow the attribution kind
Speaker:of data can lead to really kind
of impactful in a negative way
Speaker:business outcomes because the attribution
view just does not take into account
Speaker:what would've happened
without the presence of those
ads like Incre Ality does.
Speaker:And so can definitely lead brands
astray as they're looking to cut.
Speaker:Yeah, really interesting. And yeah,
Speaker:max notorious for leaning into
remarketing or branded search.
Speaker:If you're not diligent about that, it
can lean into both of those things.
Speaker:And so got to be mindful of that.
Speaker:You also quoted something
that totally ties into this.
Speaker:It's from a shop talk talk that
you went to shop Talk the show,
Speaker:and I can't remember who said
it, but if you see high roas,
Speaker:I know something is wrong and that the
auto targeting is just finding existing
Speaker:customers. Do you remember actually
who said that and unpack a little bit?
Speaker:Yeah, I forget his name and I could
look real quick. He worked for.
Speaker:Mic.
Speaker:The Post Dan Danone, the big CPG.
Speaker:Yeah, I just really appreciated
that quote because I
Speaker:mean always wonder if I live in sort of
a bubble of being super passionate about
Speaker:incrementality versus attributed metrics,
Speaker:but that was just really refreshing to
hear because I don't think that's the
Speaker:natural.
Speaker:It's not.
Speaker:Thought in people's.
Speaker:Head spend more.
Speaker:But I really think it should
kind of spark some skepticism,
Speaker:especially when your goal really
is to try to drive new customers.
Speaker:My first,
Speaker:especially if you think about both
incrementality in the context of a SC
Speaker:or pex that's blending retargeting
and prospecting by default
Speaker:and knowing diminishing returns
Speaker:are my first dollars, yes, they're
going to be the most effective,
Speaker:but if they are focused on people that
are already buying from me and my goal in
Speaker:my head is new customers,
Speaker:I should be shocked that I can
spend a hundred dollars and drive
Speaker:this amazing new customer revenue
Speaker:and not think that something is up or
even over time as I continue to spend
Speaker:our BS meters should probably
go up a little bit more.
Speaker:And I don't think they do by default. So
I found that comment really refreshing.
Speaker:Yeah, I think that
really illustrates that,
Speaker:right where it's like most of us would
think, oh, ROAS is going up great,
Speaker:we're printing money.
Speaker:Whereas maybe you should say BS
detector, something's wrong here.
Speaker:This campaigns leaning into customers
that we're going to buy anyway.
Speaker:And I'll give two examples here to
illustrate this a little bit more.
Speaker:And I'll also, since we've been
picking on branded search so much,
Speaker:I'll share a couple of ways I
think we should use it. One.
Speaker:If.
Speaker:Other competitors are
aggressively bidding on,
Speaker:just know that if you're not Nike and
you're not Adidas and you're not like Ford
Speaker:or something, it's not a
lock. If it's a new customer,
Speaker:they could be swayed by a competitor.
Speaker:And that's generally how we
like to separate it out is like,
Speaker:let's have branded search for returning
customers and let's make that crazy
Speaker:efficient or just turn it off altogether.
Speaker:If.
Speaker:It's a new customer, then again,
we want it to be very efficient,
Speaker:but maybe we want it on because we
don't want our competitor to come in and
Speaker:swipe us to give and swipe our
customer. And so one example of this,
Speaker:I did a podcast with Brian Porter,
he's the co-founder of Simple, modern,
Speaker:great Drinkware brand has become a friend
and they did a study incrementality
Speaker:study and they found, I'll
get these numbers off,
Speaker:but it was like branded
search was 10% incremental.
Speaker:So basically what that means is if it
shows that I got a hundred new customers
Speaker:from Branded Search,
Speaker:I probably would've gotten 90 of
those if I had shut it off, right?
Speaker:Only 10% were incremental.
Speaker:So then what you would need to do there
is you need a 10 x row as on branded
Speaker:search for it to even make
sense. If it's below that,
Speaker:you're completely wasting
money. Pair that with,
Speaker:and you and I were commenting
on the House analytics, HAUS,
Speaker:Olivia Corey and team did 190
incrementality studies involving
Speaker:YouTube and they showed with
tremendous amounts of rigor
Speaker:that hey,
Speaker:YouTube is probably 342 times more
Speaker:incremental, meaning if
you see a one in platform,
Speaker:it's actually like a 3 42 in
terms of incremental impact.
Speaker:And so wildly different
between those two. But again,
Speaker:we're just so drawn to in platform
row as man, we'll just say spin,
Speaker:spin spend on p max and branded search
when really we should be saying,
Speaker:let me lean into YouTube or let
me lean into top of funnel meta.
Speaker:I think both those examples
too are really good examples.
Speaker:To me it also speaks
though to the importance of
Speaker:cost per incremental almost being
more important than incremental
Speaker:percent incremental. And that's something
I always use with branded search.
Speaker:I think you and I have a very similar
feeling around branded search.
Speaker:There's definitely a
time and a place for it,
Speaker:and it's one of those things where
it might not matter that it's 10%
Speaker:incremental, 10% incremental relative
to what Google's attributing.
Speaker:If your attributed CPA
is a dollar and now it's
Speaker:$10,
Speaker:but your margin when you sell a
product is a thousand dollars like
Speaker:hammer that all day long,
Speaker:that cost per incremental is still
extremely profitable and valuable.
Speaker:And same with the YouTube piece.
Speaker:If YouTube was four times as
incremental as Google said,
Speaker:but your YouTube was crazy expensive,
Speaker:it still might not be worth it
even though it's four times.
Speaker:More.
Speaker:Incremental than the platform was making.
Speaker:And that's how I think a lot
about this with connected tv where
Speaker:connected TV can be super powerful
and maybe more so than linear tv,
Speaker:but if you can buy scatter
linear TV for a 10th
Speaker:of the cost of CTV,
Speaker:well it just has to be more
than a 10th as effective and
Speaker:it's accreted, it's a positive.
Speaker:So it becomes more of comparison
of a cost per than just a
Speaker:blanket.
Speaker:How incremental is something which I
always think is important to focus on and
Speaker:call out.
Speaker:To. Yeah, it's so good.
Speaker:I mean measuring something in terms of
percentages can provide insights and help
Speaker:make decisions, but ultimately
it's the cost per right.
Speaker:Translate that into real dollars
to see if it makes sense.
Speaker:100% agree with you,
Speaker:but I think this also goes back
to and use your linear TV example,
Speaker:and I still love TV and
connected TV and stuff. Again,
Speaker:I'll use YouTube just because
I've got the numbers in my brain,
Speaker:but with YouTube sometimes
we'll see a $5 CPM or a
Speaker:$7 CPM in certain audiences
compared to other channels that are
Speaker:15, 20, 30, 50, whatever.
Totally. And I'm like, well,
Speaker:if we're reaching the right person
and if the message and offer are
Speaker:good, how could this not work? And it's
one of those things where it's like,
Speaker:okay, we're either one of those is
off, we're talking to the wrong person,
Speaker:that's the wrong message,
Speaker:or we're just not measuring it properly
and that's where we need to look at it.
Speaker:So did you have a thought on that?
Speaker:You another question on
MM here in just a second.
Speaker:Yeah, yeah, totally. But it
made me think of the idea of,
Speaker:I think the reason I'm starting to become
way more bullish on any channel that's
Speaker:historically been hard to measure
where I think there's that arbitrage
Speaker:opportunity of costs are still relatively
low because people haven't all moved
Speaker:in because it's easy to attribute.
Speaker:It'll be really interesting
with a house example,
Speaker:does that inspire a lot
more YouTube buyers?
Speaker:That's something that Google
should have put out way long ago,
Speaker:but I think it would undermine
undermine search and that's their bigger
Speaker:business. And I could do a whole
kind of rant and I'll save you that,
Speaker:but the idea of incrementality first
measurement probably wouldn't be great for
Speaker:the search business. So probably exactly,
Speaker:haven't been able to make such a
good point that case on YouTube.
Speaker:But you think about all the channels
that have historically been harder to
Speaker:attribute,
Speaker:that's where costs are deflated just
from a supply and demand perspective.
Speaker:So when you can move in and get CPMs at
five to $7 and it's really effective,
Speaker:but most people that are measuring
through attribution don't know it's really
Speaker:effective, that's a huge win for certain
period of time until everybody's flood,
Speaker:everybody and the costs go.
Speaker:Up the market.
Speaker:I'm sure there's a lot of people that
were not excited to see that study from
Speaker:house like dang it, that means my costs
are going up. I don't like that at all.
Speaker:So really good man.
Speaker:So we talked about incrementality testing
and I think you can use tools like
Speaker:House and then there are others.
Speaker:We're just talking about work magic and
there's a number of others you can lean
Speaker:into. Full disclosure,
they're pretty expensive,
Speaker:but you can also do stuff on your own too.
Speaker:If you've got someone that
can measure this stuff,
Speaker:you can do a little bit of it on your
own. What about the MMM side of things?
Speaker:What's kind of the easy way to start
there? Is there an easy way to start?
Speaker:What do you recommend to people.
Speaker:There? I don't know. I dunno if
there's an easy way to do anything.
Speaker:I think, well, I guess
that's not totally true.
Speaker:I think there's some ways to
run relatively easy incre tests.
Speaker:So I think that's the
easier place to start.
Speaker:Certainly you can always
ratchet up the scientific rigor.
Speaker:I think the problem with looking
for an easy MM solution is
Speaker:anybody could run a model with Robin or
there's a lot of open source packages,
Speaker:but just because you can run a model,
Speaker:it could say anything.
Speaker:It's not necessarily rooted in this
can all of a sudden predict the future
Speaker:and tell you exactly the
contribution from media.
Speaker:Whereas incrementality can do
that a little more out of the box.
Speaker:You may have wildly wide
confidence intervals,
Speaker:but it answers the question.
It gives you the comparison.
Speaker:I didn't do it in this market,
Speaker:I did it in this market.
What is the Delta Media mix modeling?
Speaker:You could build a model
to tell sort of any story.
Speaker:The proof is sort of in the pudding of
if I do the thing that the model says,
Speaker:does it change my top line?
Speaker:Can I see over time that
when I listen to the model
Speaker:that improves my top line?
Speaker:So it's a lot easier to get started
with incrementality testing.
Speaker:You can run poor man's match
market tests as I sort you can just
Speaker:sort of pick,
Speaker:some markets historically behave
similarly and there's certainly some risk
Speaker:there, but with a model you might
think that it's an amazing model.
Speaker:I just don't feel like there's a great
place to DIY that together without some
Speaker:real scientific or statistical
rigor. Or if you do,
Speaker:you've just got to try to prove it over
and over by taking some big swings. And
Speaker:that's really,
Speaker:I sort of feel like you can get away
with the kind of feel it sort of tests
Speaker:without really running a true
incrementality test or model.
Speaker:If you're a small enough business and
you spend a decent amount on Facebook,
Speaker:maybe you're not willing
to turn off Facebook,
Speaker:but are you willing to drastically
increase spend and see if you can feel
Speaker:something at the top line? Okay, then
what happens if you cut it in half?
Speaker:What happens?
Speaker:And start to understand those curves on
your own is probably a less risky way
Speaker:than trying to, I've never done
anything in R and I'm going to run
Speaker:or done any sort of medium amount.
I'm going to try to run one.
Speaker:That's probably a risky proposition.
Speaker:Yeah, it's a really good insight. I'm
glad you answered the question that way.
Speaker:I think, yeah,
Speaker:leaning into the poor man's incrementality
test or just leaning really heavily
Speaker:into a channel and measuring your top
line if you've got a small enough business
Speaker:to look at that, but probably if
you're going to lean into MM M1,
Speaker:you need a couple years of data and so
to be able to make some correlations and
Speaker:you probably need to lean in to
someone or a tool with quite a bit of
Speaker:experience because you can do that astray.
Speaker:And on your comment on cost too.
Speaker:I mean it's all relative and a lot of
times where you're going to need a medium
Speaker:mix modeling is when you're spending
a significant amount in a significant
Speaker:number of channels,
Speaker:which you're probably only doing
if you are spending a lot total,
Speaker:which you're probably only doing if your
revenue can support that high level of
Speaker:spend,
Speaker:which means that a tool may not be
all that expensive relative to the
Speaker:opportunity you could derive from
it, which is where I always net out.
Speaker:So I'm paying 10 or 20
grand for a tool monthly,
Speaker:but it's allowing me to
redeploy millions in ad spend.
Speaker:And it totally in completely
makes sense. So Tom,
Speaker:this has been fantastic.
I'm just watching the clock.
Speaker:I know we're kind of coming
up against it, but one,
Speaker:I recommend people follow you on LinkedIn.
You put out some awesome content.
Speaker:I love reading it.
Speaker:Thank.
Speaker:You. People should definitely follow
you on LinkedIn and you are, is it Tom,
Speaker:what is your handle on LinkedIn?
You are Thomas B. Leonard.
Speaker:Thomas B. Leonard. That's
probably confusing.
Speaker:I'm very self-conscious of LinkedIn, so
I'm glad to thank you for saying that.
Speaker:I think it's good, man. I think it's
really good. I like it a lot. Yeah.
Speaker:Yeah, it's been fun to start
doing connecting with folks.
Speaker:Definitely an area that had a lot
of excitement and passion for,
Speaker:it's fun to have these
sort of conversations,
Speaker:so I appreciate you reaching out a
while ago and that we could connect.
Speaker:Absolutely.
Speaker:Man. Absolutely. So then if
other people were like, Hey,
Speaker:I just want to talk to Tom because maybe
you can help my brand or my business,
Speaker:how can they connect with you and who are
you looking to or who do you feel like
Speaker:you can help?
Speaker:Yeah, definitely appreciate that.
Yeah, reach out on LinkedIn.
Speaker:I spend time there. I love reading
everybody's thoughts and content. So yeah,
Speaker:reach out on LinkedIn mostly we work
with consumer facing brands that
Speaker:are trying to understand where to
put the next dollar or where to pull
Speaker:in the scenarios. They have to really
kind of rescue people from attribution,
Speaker:trying to better understand where they
can get more with their ad dollars.
Speaker:I think to your point that you teed
up now is such an interesting time or
Speaker:anytime that there's margin pressure,
Speaker:there's more scrutiny
on a marketing budget.
Speaker:Really want to try to help
empower marketing teams to
feel more confident with
Speaker:what they're doing and ultimately the
finance teams to feel more confident with
Speaker:what marketing team is doing. Hundred
percent. That's where I love to plug in,
Speaker:but also just love to talk about this
stuff probably more than I should.
Speaker:So always open to the conversation.
Speaker:Yeah, I talk about that a lot.
Speaker:I've read analytics and measurement
books on vacation and my wife
Speaker:is like, what is wrong with you? And I'm
like, it's interesting. I don't know.
Speaker:I like it. And so totally, we are
just a different breed I suppose,
Speaker:but I love that.
Speaker:And then I think this is a great way to
end it where if I've got an extra dollar
Speaker:to spend on marketing, where do I put
it? If I need to cut a dollar of spend,
Speaker:where do I cut it from?
Speaker:And that's really what
this approach is about MMM
Speaker:and incrementality. And so
I think their necessities,
Speaker:I think attribution is broken and or
misleading in so many different ways.
Speaker:There's some correlations there, so we
don't have to throw it out completely,
Speaker:but I do believe you need to lean
into MMM and incrementality for short.
Speaker:So connect with Tom on LinkedIn.
And with that, we'll wrap.
Speaker:Tom's been fantastic. Thanks for the
time, the insights and the energy. Yeah.
Speaker:Thanks so much Brett
time. Glad to connect.
Speaker:Absolutely. And as always, thank you for
tuning in. We'd love to hear from you.
Speaker:If you found this episode helpful,
Speaker:someone else in the D two C space or
marketing space, and you think, man,
Speaker:they got to listen to this, please
share it. We mean the world to me.
Speaker:And with that, until next
time, thank you for listening.