All Episodes

March 18, 2025 61 mins

Why? Or… y? What is y? Why, it's mx + b! It's the formula for a line, which is just a hop, a skip, and an error term away from the formula for a linear regression! On the one hand, it couldn't be simpler. On the other hand, it's a broad and deep topic. You've got your parameters, your feature engineering, your regularization, the risks of flawed assumptions and multicollinearity and overfitting, the distinction between inference and prediction... and that's just a warm-up! What variables would you expect to be significant in a model aimed at predicting how engaging an episode will be? Presumably, guest quality would top your list! It topped ours, which is why we asked past guest Chelsea Parlett-Pelleriti from Recast to return for an exploration of the topic! Our model crushed it. For complete show notes, including links to items mentioned in this episode and a transcript of the show, visit the show page.

Mark as Played
Transcript

Episode Transcript

Available transcripts are automatically generated. Complete accuracy is not guaranteed.
(00:05):
Welcome to the Analytics Power Hour. Analytics topics covered conversationally
and sometimes with explicit language. Hi, everyone. Welcome to the Analytics
Power Hour. This is episode number 267. I'm Tim Wilson from Facts &
Feelings, and according to the logistic regression that I ran personally

(00:26):
on the last 40 episodes of this show, there is a 72.3%
chance that I'm joined for this episode by Julie Hoyer from Further.
Julie, is that you? Hey, look at that. Yes, it is.
Here I am. Look at my model kicking ass. Sweet.
So it also says there's a 61.4% chance that Michael Helbling from Stacked

(00:50):
analytics will be another co host. Michael? Michael. Michael. Okay. No Michael.
Let's see. Next up, the model said there was a 41.7%
chance that Moe Kiss from Canva would be co hosting. Moe, are you
there? I am. But is your model any good? Well, I plugged in

(01:15):
everything I had and when I created the regression, but it still couldn't
perfectly predict who would be co hosting. Does that mean my model was
wrong? Did a model even really exist? Well, maybe the answer is I
don't think so for either one. But I have questions. And when we
have questions and we have a podcast, we get to find someone to

(01:35):
answer them. In this case, we reached back into our archives for one
of our favorite past guests. Chelsea Parlett Pelleriti, also known as the
Chartistician, is a statistician and data scientist who was our guest way
back on episode number 149. By day, she is a consulting statistician with
our friends at Recast, but she also has a passion for teaching,

(01:58):
bringing interest and excitement about math and statistics to the masses
in fun and engaging and even endearing ways. She has a PhD in
Computational and data sciences, which the last time she was on,
she was still working towards. So she has since completed that and she
was an assistant professor at Chapman University up until last year,
teaching computer science, data science, and machine learning, which made

(02:21):
for some pretty awesome social media content. She's still keeping her foot
in teaching. She's actually currently teaching a math through video games
seminar as an adjunct professor. And she just likes teaching stuff.
And maybe I botched my intro. I was doing so well.
But today she is our guest. Welcome back to the show, Chelsea. 0:02:43.7Chelsea

Parlett Pelleriti (02:42):
Thank you. It is a pleasure to be here.
But it makes me feel very old thinking how long ago
it was that I was last on the show. Okay. Well... So now
you're making me feel super, super Old. Yeah, I believe you're right.
I was in my 40s, and that was a long time ago.

(03:04):
I was in my 20s. Oh, okay. Okay. Well, the passage of time.
So this show is... It's actually a direct result of the listener survey
we did last year, which we had a bonus episode that came out
a little while back that talked about that. And we had multiple respondents
who requested in one way or another that we cover specific statistical methods

(03:28):
on the show. And this is really kind of our first attempt at
doing that. So we'll see how it goes. I'm not ashamed to say
I got pretty excited as I was thinking about this show because I
realized how much I've been faking various things for a while.
And this is my opportunity to ask questions as though I know the
answer when I don't. And then I will know. And I can ask

(03:50):
the questions like, I don't know the answer when I really don't know
the answer. That's good. Good compliment. And Julie will be the only one
who understands the answers. So there we go. We've got the full.
The full set. It'll be a refresh for me, too. I don't get
to do as many regressions in my day to day as I would
like. Well, what seemed like a great place to start would be with
that kind of absolute workhorse of prediction, which is plain old regression.

(04:16):
And Chelsea, you're pretty deep and if I understand
all the content I read from Recast pretty well, then you're pretty deep
in the world of kind of Bayesian statistics and causal inference when it
comes to doing media mixed modeling work. So does regression come up in
your day to day at all? Or is that too basic?

(04:37):
You've moved on to fancier things? I mean, there's definitely a time and
place for fancier methods. But linear regression is probably the first thing
I try in any problem that it might be a good fit for
and definitely has a place in my day to day still.
And I think there's a sense in which you can think even really

(04:58):
complicated MMM Models like you can build with Recast or other tools.
There's a sense in which it's just an extension of ideas that are
present in linear regression. So even if you're not actually using a linear
regression, you're really capitalizing on the ideas that using linear regression
teaches you. So in that sense, it never goes away.

(05:20):
That checks. I feel like I've watched my... I'll count myself as one
of the people who, when they finally understood kind of what MMM was
and then decided to try to explain it. You always wind up with
the slide that shows the formula for regression and says, look,
so your dependent variable is. And your independent variable. And the coefficients
mean. So you're saying that is a... Even Kevin Hartman has a video

(05:44):
on the basics of regression. And I think he uses, basically,
he didn't say mixed modeling, but he uses that as an example. Okay, that's
good to know. So should we define regression and we'll see where it
goes from there? Like, if you're... Did you have to do that with

(06:05):
students say, here's Regression Intro 101. How would you explain it?
Yeah, absolutely. I mean, if we're just talking about linear regression,
it's basically a model that you can look at both predictably,
so trying to actually make predictions with it, or inferentially trying
to understand the relationship between variables. And it's a super simple

(06:27):
model because all it is is that equation for a line that you
learned back in, I don't know, middle school or whenever that came up
for you in the math curriculum, where it's y=mx+b. Right? That's the definition
of an equation of a line. And that's exactly what linear regression is.
It uses various variables in order to predict something that you're interested

(06:49):
in, whether it's revenue or conversions or something like that. By combining
the predictors that you have in a linear way, all that means is
that every single predictor variable you have, you're going to multiply
it by a number, add all of those together, and that's going to
be the prediction of your model. Can we go back to when you

(07:11):
said it's still such a big part of your day
for problems where it's appropriate? When you are sitting at your desk doing
your work, what are the problems that you're like, no, I'm definitely not
trying that first. That's a really good question. I think it depends a
bit on what tools you have available. My way of working is that

(07:35):
anytime a problem comes up, I want to try the simplest method possible
to solve that problem. So if it's a problem where perhaps just a
graph is going to solve the problem and answer my question,
I wouldn't go as far as linear regression to answer that question.
And on the flip side, if I have a problem that is super

(07:56):
complicated and I know, for instance, it's going to violate some of the
assumptions of linear regression, then I might skip over it just because
I know that any answer that I get out of it might not
actually be usable and I don't want to invest my time there.
For instance, if we have a really complicated problem, for instance,
in marketing, one of the things that we talk about a lot is

(08:18):
that we have not a problem. We have an interesting scenario where when
we spend money it doesn't have an effect right away all the time.
Sometimes it takes a while for that spend to actually have an impact
in your market. And that might be something that's kind of hard to
represent with a linear regression. You may at that point need to graduate
to more complicated models. So it's say if I already know that something

(08:41):
simpler is going to work or I know for sure that my problem
just doesn't fit the parameters of a linear regression. Those are the times
when I would say, eh, I'm not even going to try.
But pretty much any other time I want to try it because it's
a super interpretable model, it's super easy to run compared to other options
and most people have a little bit of knowledge about what a linear

(09:03):
regression is. So it's really easy to communicate results even to non technical
stakeholders. So I feel like you skipped. I mean you kind of briefly
hit it as to whether or not it wouldn't be a fit.
Like using the Y=mX+B. Like you have to have a Y and that
Y has to be singular. So if somebody said, can you look at
this data and kind of put them into logical groups? Like then where

(09:28):
you don't have a, like you need to kind of have a dependent
variable and one or more like the framing of the problem needs to
kind of lend itself to saying I'm trying to find a relationship between
one or more of these things in this other singular thing.
Right. Like that's kind of the first deal. Yeah, well you can always,
yeah, you could always have like a multivariate regression where you have,

(09:49):
you're actually predicting a vector of values but in its most simple form,
exactly. And one of the things you mentioned kind of made me think
about. Yeah, it's, it's really only for supervised questions. Supervised
in the machine learning world, meaning we already know the answer for some
set subset of our data that we can train on. So for instance,
if you're doing customer segmentation, that's an example of something where

(10:14):
you don't have a correct answer. We don't know what these latent groups
are that are in our customer data. And so we can't use something,
even a modified version of Linear regression, because that's an unsupervised
problem. We don't have the answer in order to train our model.
So that's an example of where linear regression, or any extension of it,

(10:34):
wouldn't be a good fit. And you also mentioned assumptions where you said,
I have to make sure that no assumptions around a linear regression would
be broken. And then I say, yes, it's a good fit to use.
Can we talk a little bit about those? Like, obviously you just said
supervised, but what are some of the other ones? Yeah, there's a ton
of them. Some of them have really funny names, so I'll give you

(10:56):
a little bit of a warning before we get there. But one of
the most important ones is something that sounds so silly, but is the
assumption that the relationship between our variables and our outcome is
linear in the parameters. So what that means is whatever columns of your
data that you're going to plug in as predictors into your model,

(11:17):
the relationship between them and your outcome has to be linear and additive,
meaning that as your predictor variable increases, there's some type of
constant relationship between that increase and the increase in the predicted
outcome that you're trying to predict. So, for instance, if you're trying
to predict how much revenue am I going to get depending on how

(11:40):
much I spend in Facebook? In a linear regression model, your coefficient
says no matter how much I spend, every increased dollar that I spend
in Facebook is going to increase my predicted revenue the exact same amount,
not maybe the most good assumption. In a lot of cases,

(12:01):
an additive basically means that all of the impacts that my different predictor
variables are going to have, they're kind of independent, and we're adding
them all together at the end. And that's just reflected in that Y=mX+B formula,
right? Every single predictor is getting multiplied by a constant coefficient.
And then we're adding up all of those effects to together to get

(12:22):
our predicted outcome. But that doesn't mean the predictors can't be squared
or cubed or combined across multiple interaction effects, or basically taking
X1 and an X2 and multiplying them together. Exactly. Okay, okay,
now you've got to go a bit slower. You got to go a
bit slower on that one. You lost me. Okay, that's linear in the

(12:43):
parameters, right? The parameters of your regression model are the intercept
and a coefficient or multiple coefficients. So when we say linear in the
parameters, all we mean is that whatever it is that our predictors are,
we're multiplying them by a constant adding them together, but exactly as
you said. So say you think that there isn't a perfectly linear relationship

(13:06):
between... We'll stick with the example of Facebook spend and revenue.
One of the things that we could do is I like to call
just feature engineering. We could take the amount we're spending in Facebook
and we could add a new column that is the amount we spent
in Facebook squared. And so now we have two predictors. We have how
much we spent in Facebook, how much we spent in Facebook squared.

(13:30):
But that still fits into the mindset of linear regression because the actual
columns you're plugging into your model still linear in the parameters.
Our Facebook spend squared is just a new predictor that we are assigning
a coefficient, multiplying it by that constant coefficient, adding it to
our prediction. And the same happens for interaction. So interaction terms

(13:52):
are just the value of two or more predictors multiplied together.
And again, it's just feature engineering. You're creating a new column to
put into your regression model and that allows you to understand the relationship
of when these two things co occur together.
How am I adjusting my expectation or my prediction based on the fact

(14:14):
that they're occurring together? And I think they have some type of relationship.
So there are a couple ways. The feature engineering thing is kind of
a fascinating. And I have, I think, two questions. One choosing the features
feels like there's a lot of kind of art in that.

(14:35):
But there's also, as I understand it, there's a risk that if you,
if you chose two features that are strongly, strongly correlated with each
other, it could cause problems. And why am I now blanking on what
that is? Multicollinearity. Multicollinearity. And it was there. I was like,

(14:56):
yeah, it's there. Yeah totally there. I was the tip of my tongue.
It was. But I guess so that's because if we look at it
and say, I'll just throw it at the data, but the data is
just going to have X, it's not going to have X squared.
It's not going to have the square root of X. It's not going
to have x1 meets x2. What's kind of the

(15:19):
approach and what are the risks if you try to get too fancy
with that? Well, one of the risks is that you'll be just too
fancy and no one will want to talk to you.
But the major risk there is that you're going to be misspecifying your
model and/or you're going to be overfitting your model. So this is a

(15:41):
super common thing. If you're not really thinking through what it is that
you want to include in your model. And let me give you an
example. So let's say that I am fitting. We call... So when we
add X squared, X cubed, blah blah blah to the predictors of our
model, we usually call that polynomial regression. And if you have a polynomial

(16:02):
of 75 degrees, I don't know if that means anything to you,
but what that means is that the line that you can fit to
your data is incredibly wiggly. It can literally hit every single point,
probably, or a lot of the points in your data set.
And that's going to lead to overfitting. So one of the risks that
you take when you do things like this feature engineering is that you

(16:24):
might be overfitting. Now, there's ways around that. I don't know if you
want to get into things like regularization which can help you understand,
or like pull back the impact of terms that are... Is that where K fold cross
validation would fit in. No, damn it. Not quite. Okay. Very related though.
Yeah. Too soon. I think you're jumping the gun a little bit.

(16:45):
Okay, but, but the other point I wanted to make there, so you
have tools like regularization, I'll give you... So the real world way to
define regularization is it's any method that makes your model a little
bit simpler. In the practice of regularizing coefficient estimates, usually
what that means is pulling them closer to zero, unless there's evidence

(17:06):
that having a non zero coefficient really improves the fit of your model.
So if you're familiar out there with lasso or ridge regression,
that's what they're doing. Those are methods of regularization that basically
encode the idea that in the real world most effects are exactly zero
or close to zero, unless we have pretty strong evidence to the contrary.

(17:29):
So that's one thing you can do. But my favorite way of approaching
this is always leveraging the subject matter expertise that we have. I have
a really technical background, but I can't build good models unless I understand
the context of what we're building and what we're building for.
And so honestly, my favorite way to approach this is often to get

(17:51):
the opinion of subject matter experts who are able to give insight,
at least to some degree about what we should include. For instance,
we are talking about interaction terms which basically says when two predictors
co occur together, how does that change our prediction? A lot of times
subject matter experts will have a good idea about which ones should be

(18:13):
included in the model and which ones are just so ridiculous,
we would never even want to try them. So while there are statistical
techniques to handle the risks there, my favorite way is with,
if it's available expertise. How do you go selling that to especially,
like clients and whatnot? Because I feel like sometimes they think it's

(18:33):
like disconnected. Like you go away and build the best model and
sometimes convincing them that you need the business context or I guess,
guidance on assumptions. I don't know, I feel like there could be resistance
or that there could be pushback, that they're bringing bias in.
Like, what's your take on that? Oh, my gosh, that is a huge

(18:55):
concern. Especially I work a lot with Bayesian models. So you not only
have kind of this prior information coming through the settings you're choosing
of your model, what type of model you run, but also through the
priors that enter into the analysis. And I think the way that I
talk about it has to do with leveraging different sources of data.

(19:16):
Right. So we actually have your data, the data that we're going to
use to train your model to fit your model. But we also have
sources of information that come through your years or decades of expertise.
One of the examples I liked to give in my classes,
even before I was in the marketing space, is that if I'm doing
an experiment and I'm looking at the click rate of an email that

(19:38):
I'm going to send out, maybe I'm like A/B testing it.
It would be silly of me if I didn't leverage the expertise of
the marketing people in the room who know that click rates are probably
going to be, I'm making this number up around 2%. It would be
insane if I got an 80% click, like something went wrong there if

(19:59):
I got an 80% click rate. And while I understand the desire to
not bias your model, and certainly that can be a concern,
it really feels like throwing away information to not include that expertise
in your analysis where appropriate. So through a Bayesian prior or through
some model setting configurations, another thing that comes up especially

(20:22):
with MMM, these are such complicated models. Not always. I think Tim might
have a story about not the most complicated MMM model, but basically these
are often really complicated models and there just isn't enough data for
us to get really good estimates on the parameters that we would like
to know in your model. And so we need that prior information in

(20:44):
order to fit a model that makes sense. Right. If you tell us,
hey, ROIs are certainly not going to be 200,
that's really helpful to the model. And I don't think that's biasing.
But at the same time, we do caution people a lot to make
sure that they're not overly specifying the prior information in a way that

(21:05):
tells the model exactly what to say. We still want the data to
have a say in what the model is learning. But it would be
silly to assume that you don't have expertise from years and years in
the marketing space that can help us inform our analyses. But I will
say there is a lot of pushback. And often what it comes down
to is we'll put very loose parameters of say, okay, if you really

(21:29):
don't know, here's different things that you can try and we'll show them.
When you have information that's reasonable, it actually improves the insights
you can make. And that really happened to me a lot when I
was a professor. We would often do a lot of consulting with people
who weren't super familiar with Bayesian models, for example, and there
was so much pushback about I don't know what a prior should be.

(21:51):
Shouldn't the data tell me what the prior should be? And what I
usually like to tell them, at least in the psychology space,
which was primarily where I was working, is, okay, if I told you
that this intervention that you're testing had an impact where it improved
people's IQ by 70 points, what would you come back and tell me?

(22:14):
And most of the time they would say, that's insane. You did something
wrong. And I say, that's prior information, even if it's very loose.
You're giving us a little bit of an idea of what a reasonable
value to expect would be. And that can be super helpful in an
analysis. Tim, I'm drawing so many parallels right now between this and
when you talk about bracketing for setting targets. What's your reaction

(22:38):
to that? Yeah. I mean, I was thinking the same thing.
Like, I will frame. I mean, the same thing. Running into setting targets
for KPIs, and people say, what are you talking about? Like,
I have no idea. And then you say,
well, what if your KPI is to give them this drug or this
intervention and increase their IQ by 70 points? And they'd be like,

(22:59):
well, no, like there's no way it's going to do that.
I'm like, oh, well, I guess you do have some expectations. But I
had not made the link to using that for as priors to go
into a Bayesian. That's wild. Does it work, if you've got,
say, two parties, they're both subject matter experts, and one, we'll just

(23:23):
go back to Facebook and say, one says, I think Facebook is definitely
heavy is driving sales. And the other person says, I don't think Facebook
is doing anything. Is that still useful to say? Yeah, we'll plug it
in to... We'll include it in the model. And then. And then the
model can, with caveats, come back and say, yeah, it looks like there

(23:50):
is a detective now whether it's actually Facebook or something. Confounding
separate discussion. But is that still useful? It's like this could be at
play included in the model, and then see if it's detectable.
Yeah, I mean, in cases like that, where there's not a lot of
overlap in the subject matter experts opinions, you have a couple ways that

(24:15):
you can approach this. One thing that I would do before I even
ran a model, though, is ask them why they think that is their
internal data. What is the assumption that they're making that made them
come to that conclusion? Because I would be
super shocked if two people with access to the same information,
with the same assumptions came to such different conclusions. We never want

(24:37):
to include vibes as priors. We want to have kind of informed decisions
and reasons why we believe this stuff. And so that would be my
first line of defense, because I think once you uncovered that,
you'd figure out that they're making really different assumptions, they're
applying it in really different contexts, and that's why they're coming
to different answers. But if magically that was the case, that they really

(25:00):
had the same information and they were just coming to very different conclusions,
I think one of the things that you could do is you could
have a very wide prior. Right. When you take the collective expertise of
the different experts, then there does seem to be a lot of uncertainty.
And one of the things that you said that was interesting is you're

(25:21):
kind of talking about the point estimate of what they expect in terms
of, let's say, an ROI. One thinks it's really low, one thinks it's
really high. But when you extract information that might be useful in an
analysis as a prior, you really have to make sure that you're also
thinking about uncertainty. So maybe if you ask them for uncertainty around

(25:41):
their estimates, you would find a lot more overlap in what they believe
and it might be easier to translate. That being said, if they don't
and they can't agree, then I might do something called a sensitivity analysis
where I ran the model with expert one's priors and ran the model
with expert two's priors and then saw how the model performs.
Right. If your model can't form forecast well, if it has high variance,

(26:03):
it's adjusting all the time and giving you different insights. That's a
bad model. And if one of the priors lends itself more to that
poor performance, probably not a good prior, or maybe there's something
misspecified in our model. But I would test something like that and just
see like, does it negatively impact the insights we're getting from the
analysis? Can I take us back a little bit? When you talked about

(26:26):
using regression for prediction or like inferentially, because when you're
talking about like this feature engineering, do you have to go about it
differently when you are using it to create a model to predict something?
Right. Compared to like you're trying to inference something? Because I

(26:46):
think I've used regressions more. So when I'm looking at historical data,
we have a business question they're asking and we're trying to infer if
a relationship exists. And I feel like that I've always read things and
had the experience talking to colleagues where you have to be really careful
about not looking at the results and then tweaking. Right. You don't want
to like bring in bias of your features to make it look good

(27:08):
at the end. And so we always had these discussions about choosing your
features and doing a lot of other work to determine what relationships you
need to represent or not before running it and kind of being like
that's the answer. Whereas it sounds like though when you're using it for
prediction, you're trying to find the model that has the best fit on
what you're training it on so then you can use it moving forward

(27:29):
with high confidence, higher confidence. Right. But can you, am I misinformed
there? Like, do you do feature engineering differently for those two scenarios?
Yeah, that's a great question. I think in general, because your goals are
so different, right. With prediction, all you care about is that the output
of the model, the thing that it's predicting is as close to what

(27:51):
the real value is in the real world. Whereas when you're doing inference,
what you care most about is, are the parameters of your model,
say your regression coefficients, are those accurate to the real life relationships?
And often that kind of veers into causal inference territory as well.
But basically I would say you would approach it differently, there's a lot

(28:11):
of overlap. But for instance, in the predictive space, yeah, we might have
a little more freedom to play with our model and try and get
the best prediction, but we have to be really still careful about overfitting
to the sample of data that we have. Which is to Tim's earlier
point, why we do things like cross validation or any type of model
validation that basically says if we hold out some data from being used

(28:36):
to fit the model, can the model still make good predictions?
Because if it can't, then it means that we've probably done what you
were describing, which is we've kind of over engineered our model to fit
fit too specifically to the sample of data that we have.
And it doesn't generalize well. And so that's the bane of every data
scientist life ever is overfitting. And so we do still have to be

(28:58):
careful about that in a predictive sense. However, when we're doing inference,
sometimes we have to be a little bit extra careful. An example that
I like to give is say you're a bank and you're trying to
predict if someone is going to default on a loan that you've given
them. In a predictive sense, if knowing whether that person has a yellow
car helps you understand and makes good out of sample predictions for whether

(29:23):
that person is going to default on their loan, I don't care.
I don't care if that's a real relationship, if that's causal.
I just want to know if this person is going to pay me
back or not. Whereas if you're doing an inferential model, I think we
might want to put some more thought into that. We want to do
a dag, right, to look at the causal relationships and see if maybe,

(29:44):
I don't know, there's some confound there between having a yellow car and
defaulting on your loan. And so you might want to put a little
bit more thought into that. So wait, have I got this right?
So in the case where you're trying to predict,
you don't care about a causal link with the yellow car.
Have I got that right? It depends. Because you care about the prediction

(30:07):
as the primary thing that's the most important, the accuracy of the prediction.
And then take me through the second piece. Yeah, so in the predictive
context, first of all, if having a yellow car doesn't help me predict
values that the model's never seen before, right. So if the out of
sample accuracy is bad, do not want to include it. But if it

(30:28):
just is say, tangentially related to a different construct we can't measure,
maybe wealth or eccentricity or something like that that we don't have a
good measure for. But having a yellow car is a proxy for,
then if that helps me make a prediction, great. It's giving me information
about something else. Even though having a yellow car itself is not what

(30:50):
is making someone default on a loan, well, maybe I don't know the
power that yellow cars have. In an inferential case, what we really care
about is, are the relationships that we're modeling accurate to the real
world. For instance, if I have, let's go the frequentist route,
if I have a significant p value on my regression coefficient for having

(31:15):
a yellow car, then is that a real relationship? Is that a causal
relationship? And when we say causal, what we really mean is if we
change the color of your car, if I got in my
Camo garb and re spray painted your car, is that suddenly going to
change how likely it is that you're going to default on your loan

(31:36):
payment? That's what we really care about. And so in that case it
does matter that really, I guess if I'm thinking about it deeply,
I don't think the color of your car is causally related.
I think there might be some other process like your wealth or how chaotic
and wild you like your car colors, that it's causing both loan default

(31:58):
probability and causing you to have a yellow car. In that case,
then I might really care. I'm not really truly estimating a causal impact
there. And so that insight isn't going to help me. Right.
If I am a bank and I want more people to pay me
back, I am not going to then go out and spray paint their
cars yellow because that's not a causal factor in paying back your loan.

(32:20):
But that's, I mean you mentioned kind of a DAG in passing that
goes, it goes back to talking to a subject matter expert who may
be into the psychology and they, if you were diagramming it out saying
why would a yellow car. And they said, oh, that's actually a... That
may be related to flightiness or eccentricity or something. You can capture

(32:40):
all of that as sort of assumptions or kind of likely relationships that
could then guide. Right, because that could give you like. Well,
instead of yellow car, if a yellow car is kind of a proxy
for something else, there might be a better proxy. So maybe you should
look at that other proxy that we can measure and use that instead
which just gets you back to picking parameters while working with a subject

(33:05):
matter expert that are the best. Yeah. You can't build very many good
models without a subject matter expert. And I think it really comes down
to, are you asking a causal question? And often in the space of
marketing, we are. Right. When we want to know the effectiveness of,
we keep saying Facebook, there are other marketing channels. But if you

(33:27):
want to know the effectiveness of Facebook, what you're really asking is,
if I change my Facebook spend, is that going to have an impact
on whatever it is that I'm measuring? And so we really are asking
a causal question in a lot of these scenarios. If you just want
to know what things are associated, then we might not care.

(33:49):
Right. So if you have a model that's predicting the LTV of a
customer, again, it might not matter if you know that they have a
yellow car, they're going to spend huge amounts of money with you.
Great. That helps me kind of forecast what my customer LTV is.
Plan accordingly for whatever server space or whatever it is you need to

(34:09):
serve customers like that. And yet it's not the case that I might
want to go spray paint my customers cars yellow, because that's not actually
going to have an impact. So often in marketing, we are asking these
causal questions because we want to be able to take an action and
understand what impact that action would have on whatever it is we're measuring.

(34:30):
Okay. Weird question. I feel like I'm always doing this and I'm like.
But about the business. I'm just curious. So I feel like I'm with
you on the yellow car analogy, which I love, by the way.
Although I do want to see you in like, some kind of guard
spray painting Tim's car at some point. Can do. My first car was

(34:51):
a yellow 78 Chevy Monza. So, you know.
Yeah. And I've always never defaulted on a loan, so. Well,
there you go. Proof. Okay, so the bit that I... I'm,
yeah, obviously much more on the business side. How important is it?
Like, those two scenarios make sense to me. One is about getting the

(35:11):
best prediction. One is about establishing causality. If I were a stakeholder,
though, and they saw a little bit under the hood of this yellow
car situation, they probably wouldn't understand why in case A, it's important,
but in case B, it isn't. So when... I mean, how much are
you showing under the hood? Or how important do you think it is

(35:34):
to understand? Tim is pointing out his book being like, Moe, go back
and read it for the fifth time. No. Buy it for your...
Get it for your business partners. Yeah, no, sorry. I would love Tim's
take on this since he pointed out his book. What do you think?
Yeah, Tim, hot seat time. Oh, shit. I just remember we were...

(35:56):
And it was Joe Sutherland who kind of made that. Point I think
in the book, more about... I think you can have that discussion.
Like, what are we really trying to do here? Like, I feel like
the way Chelsea, you were framing it do you care if the analyst
has a really good understanding of the distinction in that specific case,

(36:20):
like with the yellow car? Like, do you care more about just making
the best prediction of whether they are going to repay, default on their
loan or not? Or are you really trying to understand
kind of the relationship as to what's causing it? And I think in
business a lot of times they would say it's more about

(36:40):
prediction than... I mean, I think that's like
that is a fundamental concept that I wish business users could make that
distinction. I don't think we try to educate them. And it's not that
hard to do. Right. And I wonder too if you ask them,

(37:01):
is the data point of the prediction, their likelihood to default,
going to help you make business decisions, or is the data point that
having a yellow car has a causal relationship with you defaulting?
Like, which one's going to help you make a business decision?
I would argue it's the first one. Right. Depends where you are in

(37:22):
the business. True. But if I know you have a yellow car I
can't change your yellow car. I mean, I guess you might, treat them
differently. You could target yellow car havers. Yeah. Okay. And there you
go. That's why I think in marketing you would be more interested in
the causal relationship. In finance, you would care more about the predictive
quality because you're responsible for like a company forecast and you care

(37:44):
about accuracy. So even where you are in the business might make a
really big difference. That's a good point. And actually I'm. I have an
example that I have. I selfishly want to know if a regression then
was a good choice here because I feel like it's
maybe, maybe it's... I think it's fitting more in the inference side.
But we were working with a hotel chain and they were.

(38:08):
By the way, if you're listening to this and you're at a hotel
chain who worked with Julie, it was the other hotel chain she was
working with. This was totally not you. Yeah. Totally.
I mean, it was an interesting question. I'm more questioning if I chose
the right way to analyze it. So it's on me. They wanted to
know, they had this idea that to help them personalize search results online,

(38:31):
that they should be using distance between the person searching and where
they wanted to go as like a feature. And so what we started
to do was try to say, okay, is there actually because they were
trying to obviously increase profit, have people stay places longer spend
more with them, whatever. So we broke it down and said,
okay, well, let's start and say, is there even a relationship between distance

(38:55):
between where they're searching and wanting to book and
profitability? And we broke down profitability into like three ways to look
at it. But we used a regression and we ended up finding that
like, sure, it says there's a relationship that exists, but like,
it's not actually a good, it's like not impactful. It was like you

(39:15):
had the statistical, like p value was there, but the actual coefficient,
like we pretty much were able to tell them that no,
these are not great variables to use, possibly as like eventual predictors.
And again, it was kind of like a weird one. We worked a
lot with some co workers that it was a, it felt sticky.

(39:36):
But we were pretty much trying to just say like, does a good
relationship exist here? Or not to use this almost as like a feature
in a more complex model or prediction. Well, I think the good news
is I don't want to set this up as a dichotomy of like
inference and prediction are these two concepts and they're both worthy
goals. They're not completely separate. Right. We often want to make sure

(39:58):
that like we have a good model and that looks the same in
both concepts contexts. But in your specific case, you're actually bringing
up a really interesting thing. Are things impactful versus are things statistically
significant? And I think that that's a really important distinction because,
for instance, I might be so confident, especially in these kind of big

(40:22):
data scenarios, I might be so confident that when someone is further away,
let's say every hundred miles further away they are from their destination,
it's going to increase my profit by a hundredth of a cent.
And I'm so sure that's true. But that might not be something that
you want to action on because of the cost of implementing whatever algorithm

(40:44):
or whatever promotion or whatever it is that you're going to do in
response to this. It's just not worth a hundredth of a cent per
hundred miles. And so I think that's a really important distinction because
when we are doing statistical inference, one of the methods we can do
that with is frequentist statistics. And often what you're looking at is
a p value which just tells you how confident am I.

(41:07):
Well, let me rephrase this to be more specific, a p value basically
tells you how compatible is what we're observing with a world where this
effect truly is zero. So if there really is no effect between the
distance that someone is from their destination and whatever profit metric
you're using, how compatible is the data that I observed with that world?

(41:30):
And often the answer is going to be not compatible. This would be
a ridiculous thing to observe in a world where there is no relationship.
And that's what the p value really tells you is like how compatible
is what I observed with this idea of not having a relationship.
But that being said, even if I am pretty certain
that my data is incompatible with a world where there's no relationship,

(41:53):
it doesn't mean that that relationship is like practically significant is
often the term people use. And so I think you have to distinguish
between that. Now that being said, we're sort of veering into the territory
of using p values to do variable selection in regression models,
which is a little bit iffy and kind of a can of worms
to get into. But I do think that that distinction between a statistical

(42:16):
test and a practical significance of the result is so important.
And it comes up a lot if you're like A/B testing something and
you get a non significant result, oftentimes people just throw those tests
away, they'll say ah, not significant. Can't use any of the insights here,
but that's the wrong way to think about it because if you think

(42:39):
about the way that frequentist testing works, which is often what is reported
with these tests, you might have a non significant p value because the
null is true. You are actually living in a world where your A/B test
had no effect, whatever variant or intervention you're testing, no effect.

(42:59):
Or it could be that there's so much uncertainty about the effect that
we can't exclude zero. But there might be some evidence that the effect
is positive or negative in either direction. And by itself null hypothesis
significance test don't distinguish between those scenarios. And so it's
one of those cases where it's super important to not just use what

(43:22):
typically is outputted. So a null hypothesis significance test, maybe you
pair that with an equivalence test which tells you if something is practically
an effect is practically equivalent to zero. Maybe you look at effect sizes
to see what that effect is or how precise your measurement is.
Are we very certain it's zero, or are we certain so uncertain it
could be zero? Because those are very different things, very different results

(43:45):
that you can get from a test. Oh, this is killing me.
Because we are not going to get to logistic regression. We're not going
to get to talking more about squares and how OLS is least squares
and why is it squares. We're not going to talk about time and
why time is uniquely. And if I stop now, Moe's going to say,
I have one more question and that's why I left time for Moe.

(44:07):
One more question. I do. I have a really stupid question.
And it is something that has been top of mind since we first
started speaking. And dear listeners, I promise we will put up a picture
of this in the show notes because I can't talk about it without
the picture. And I realize we're on a podcast, so.

(44:32):
Okay, so I'm gonna do my best to explain what we're looking at.
Basically, on the left, we have a line kind of going up,
like what you would see in a typical linear regression. As you increase
spend, revenue goes up, profit goes up, et cetera. Right. And I feel
like I spent so much time looking at this particular graph,

(44:52):
and the reason I think this is really important is because I sometimes
wonder if people's familiarity with linear regression means that often we
interpret this relationship. You increase spend, revenue increases linearly.
And I sometimes get concerned that we're always trying to untrain this out

(45:16):
of our stakeholders. And so the graph that we have on the right
shows a diminishing return curve. So basically, for every extra dollar you
spend, you have less and less revenue. Right. And I feel like,
fundamentally, for a business that uses MMMs, we're constantly trying to
unpick this with our stakeholders. And I'm really curious to get Chelsea's
perspective. Like, we constantly are talking about linear regression. So

(45:39):
easy. It's so simple. It's great. Everyone should know how to use it.
And I feel like I've got the opposite problem where I'm like,
it's so common. I feel like I'm trying to get people to unlearn
it. And is that a fair observation? Yeah, I mean, I think the
problem is that we teach such a fixed set of tools.
When people learn something like linear regression, it's often not taught

(46:02):
in a way that would allow them to plug and play some more
complex methods on top of it. When really what we should be teaching
is those base core concepts of regression that then allow you to plug
in. Oh, before we plug in spend into our regression model,
we're going to saturate that spend because you can't spend a million dollars
in 10 minutes and have the last dollar be as effective as the

(46:24):
first dollar dollar. But I think that that's something that's really important
to distinguish because it's not that regression is wrong here, it's that
we also need something like that saturated spend on top of it that
makes it complicated. And because people have learned linear regression
in such a fixed way, that results in the graph on the left,

(46:44):
which for again, people listening at home is a straight line.
No matter how much you spend, every dollar is going to bring you
the same same amount of revenue as the previous dollar. Because we've taught
regression in such a fixed way, people aren't able to make that generalization
that, hey, what if I actually plugged in not spend but saturated spend,

(47:05):
that talks about, okay, when I am spending something, I am not necessarily
going to get a dollar's worth of effect from my millionth dollar.
Maybe I'll get $0.01's worth of effect from my millionth dollar.
And so if you think about it in that terms, we can still
use the ideas behind linear regression, but it does have that added complexity

(47:26):
on top. And because we teach linear regression sort of as this thing
that comes out of the box and you can't really alter it,
it's hard for people to understand that those two things are still incredibly
related. And it makes the real world scenario with saturated spend feel
much more complicated than it actually has to be. Because when I explain

(47:46):
it like that of like our predictor is not just spend,
it's saturated spend, that's something that can maybe make a little bit
more sense. Although the actual complexity of implementing that is a bit
more difficult, at least the insight is still really comprehensible. But
that's not what they're thinking about when they're thinking about linear
regression, which is probably causing a lot of the problems

(48:07):
that you're describing. But if you're talking to a subject matter expert,
I feel like you just described it in a really, really good way how
are you going to spend the millionth dollar? Well, you're gonna be casting
your net broader like you targeted initially why... So it feels like,
I don't know, that feels like another one where you're acknowledging the

(48:29):
importance of the marketer being a subject matter expertise and
that what they actually know does kind of play into
what the analyst or the statistician is doing with the data.
Like they can come together. It can be a Kumbaya moment.

(48:50):
They should. Yeah, yeah. I'd argue you're gonna do much better work that
way. I guess my thing is I feel like the marketers who are
like on the tools get that, right. Because they understand saturation.
I think it's more when you're trying to explain it to senior leadership
that it gets really tricky. Well, you can give them some examples of.

(49:12):
I always liked the example of if you have an influencer channel at
first, you're going to be able to scale up really well.
You're going to be able to find influencers that are targeting your niche
audience or people who are really likely to align with your brand.
But if you keep spending up, you're eventually going to get to that
random influencer who has 100 followers and all they do is review toe

(49:34):
socks. Is that going to be an effective spend for you?
No. And I think that that's really important that like you are really
limiting yourself. If you wanted to spend a billion dollars in influencer,
you're really not going to be able to scale up in that way.
And hopefully that's an example that makes them both laugh but also go,
oh yeah, my last dollar is not going to be effective because I

(49:56):
can only find so many influencers that are really aligned with my brand.
All right. I'm going to use the toe sock example for sure.
Toe socks and yellow cars. Toe socks and yellow cars. It was good.
I saved buffer for Moe's last question. It created chaos that you,
dear listeners, did not have to experience because of the magic of editing.

(50:20):
With that, we are going to have to wrap and the last thing
we like to do is go around and do a last call.
Something that each of us found kind of interesting, share worthy,
related to the show or not. And Chelsea, you are our guest.
Would you like to go first with the last call? I really would.

(50:43):
You said this didn't have to align perfectly, but I did want to
choose a recommendation that I think listeners who would enjoy this episode
could also use. So I actually want to recommend a YouTube channel.
I hope I'm pronouncing this correctly because I've never heard them say
it on the YouTube channel, but it's ritvikmath. And it is an excellent
resource for data, science and statistics. If you're just starting out or

(51:06):
even if you're a little bit more advanced is a way to get
really intuitive simple explanations of data science concepts. I've definitely
used this as a resource when I was a professor and honestly I
still use it now as a resource when my job is explaining statistical
models to people who are not building statistical models every day.
So highly recommend I will maybe link some of my favorites for you

(51:29):
to put in the show notes. That would be
awesome. Wow. More fun videos to watch. So we know what Michael that's
where Michael Helbling's gonna go to the video guy that he is.
All right, Moe, what about you? What's your last call? I have a
twofer, but one's kind of cheating a little bit.
Firstly, I do want to do a shout out to the Recast blog.

(51:53):
It is absolutely phenomenal. There was a recent post that I just saw
that is hitting very close to home and there's a lot of
learnings about like MMMs, incrementality, all that sort of stuff that I
just find are written so simply and normally using very similar ways of
explaining things that I probably would adopt so but probably more eloquent

(52:15):
and refined and available for you in a blog. So definitely check it
out at getrecast.com. My second one is I am in the midst of
our performance cycle where we do like all our performance reviews and growth
and impact goals and role changes and all that sort of stuff. And
unsurprisingly I've been using a lot of AI, but also I wrote a

(52:39):
prompt to look for bias and feedback, which led me down a complete
rabbit hole. And then I found this article that was really interesting which
was called Unfair but Valid Feedback. The seeming contradiction and the
thing I really took away from it is sometimes feedback feels really unfair

(53:00):
or potentially biased, but it could still be valid too. And so like
how do you take away the valid component while potentially challenging,
like the bias or the way it was delivered or something like that.
So I just, yeah, I found that one super, super interesting and made
me think really differently about the feedback that is being shared very

(53:21):
widely with both myself and across the team. Very cool. So yeah,
those are mine. Nice. Julie, what about you? Totally different spectrums.
Well, mine is going to take us for another turn, a hard turn
here. But we were talking about our pets with fun names at the
beginning of the episode. And I had recently read that an animal microchip

(53:45):
company, Save this Life, abruptly shut down recently and there could be
tens of thousands, hundreds of thousands of pets that use these Chips and
they just aren't a business anymore. And why I kind of chose this
one was one, the pet component, but two we recently have talked about
on episodes too, like what happens when a company gets bought and what

(54:08):
happens to their whole database. Well, in this case, what happens when a
company shuts down and they have a whole database of data people are
expecting to be able to use if they lose their pet,
whatever. So I wanted to call that out more as a PSA to
everyone who loves their pet and might have them chipped. You can get
it checked. They included that chips starting with 991 or 900164 are from

(54:30):
Save this Life and you can actually reregister them. And there are lots
of other companies you can re register with. And they even included that
only 6 in 10 microchips are actually registered. So I guess overall,
maybe just check that that chip is registered so you can find
your beloved pet if you ever need to. Going hard on PSA. What

(54:51):
a fuzzy way to fit. Tim, what about you? Well, just because I
think it was before we might have been recording it maybe in the
outtakes. So there's stories behind everybody's pets names, but Chelsea,
what is your dog's name? My dog's name is Nova, which is short
for Anova, or An Analysis of Variance. She's named after that because I

(55:12):
got my start in psychology statistics where ANOVAs are really big.
So a little nod to my past. Yeah. So additional cred.
If the content from the episode wasn't already like, wow, she knows her
stuff. She's got a dog named Nova. So my last call is a

(55:33):
blog post by our long time friend of the show, Matt Gershoff,
called Adjusted Power Multi Arm Bandit. And what I really liked about it
is that it's literally just working with like, a basic understanding of
power and confidence in looking at test results and how to think about

(55:54):
test design under different scenarios, which anybody who's dealt with Matt
Gershoff much is sometimes left very, very confused. And I almost followed
it all the way through, but I liked it just because it was
like, if you think about what you're really trying to do,
even if you've got 10 variations that you're trying to test,
that doesn't mean you definitionally need some crazy sample size if you're

(56:17):
really clear on what problem you're trying to solve. So maybe that even
gets to our inference/prediction discussion earlier. So it was a good,
fun read. So with that, Chelsea, thanks so much for coming back on.
I don't know how long we'll have to wait to ask you to
come do a part two on regression or hit one of the other

(56:41):
I mean, ANOVA could be the next next one up. It's basically regression.
We've essentially covered it. Wow. Okay. Boom. We got a twofer.
So you your activity on the socials, where are you hanging out the
most these days? I am hanging out mostly still on Twitter and also

(57:06):
Bluesky, which I'm not gonna lie in my head, is pronounced blue ski.
So I was really scared I was gonna say that out loud.
So I'm there. You can also find me on TikTok, though I don't
post a lot. Anytime you search for @Chelseaparlett, I should probably come
up. But yeah, I think those are the two places that I post
the most. And you'll get the most meme for your energy in those

(57:28):
places. I'm finding out how little I am on Twitter now and that
I have had the longest, slowest direct messaging going back and forth with
Evan LaPointe like, like two weeks. Because I'm like, oh, shit,
I'm really not checking Twitter. But I am, I feel like I'm almost
fully converted to Bluesky. Skeeting on the Bluesky. Don't reinforce that

(57:52):
it's gonna come out one day. So no show would be complete without
thanking Josh Crowhurst. We've. We've not given him too, too much special
work to do on this episode, but there's a couple special little challenges
for him on this one. We love to hear from our listeners so
you can find us on LinkedIn, the Measure Slack. We would love a

(58:15):
review on whatever podcast platform you listen to us on. If you want
to reach out and tell us what you thought of this.
Our first attempt at hitting a statistical method, if you were one of
the listeners who requested that or if you weren't but thought,
hey, that was a terrible idea. I mean, that was a great idea.
Would love to hear what you think. So regardless of whether you are

(58:37):
regressing logistically or linearly or whether you're inferencing or predicting,
no matter what you're doing for Moe and Julie, I know they join
me in saying you should keep analyzing. Thanks for listening. Let's keep
the conversation going with your comments, suggestions and questions on

(59:00):
Twitter at @AnalyticsHour on the web at AnalyticsHour.io, our LinkedIn group
and the Measure Chat Slack group. Music for the podcast by Josh Crowhurst.
So smart guys wanted to fit in, so they made up a term
called analytics. Analytics don't work. Do the analytics say, go for it,

(59:21):
no matter who's going for it. So if you and I were on
the field, the analytics say, go for it. It's the stupidest,
laziest, lamest thing I've ever heard for reasoning in competition.
And then there are people who just have hot sauce. They give me
another dog after. Yep, yep, we have a Cholula. Yeah, Good hot sauce.

(59:42):
Great. I remember Moe's dog's name. I can. My dog had.
My dog had her name when I got her, but I decided to
reach change the spelling. So that was Kaylee from Firefly instead of Kaylee.
That's a very obscure TV show. That's some other nerd. That's some nerd
tread. That is literally the only nerdy thing I ever have been able

(01:00:03):
to say ever. Ever. That is so not normally my thing that I
could just throw in there. But anyway, I was going through a five
minute. Oh, I got two or three ways I want to go with
this. So unless foul or foul, literally the person who's not here.

(01:00:31):
Wow, you spend way too much time with that lady.
Yep, I do. I should not have rolled off a LinkedIn Live earlier
today onto a podcast. Okay, so, yeah, like I said, we'll stop.

(01:00:59):
Rock flag. And significant doesn't mean impactful.
Advertise With Us

Popular Podcasts

Stuff You Should Know
Dateline NBC

Dateline NBC

Current and classic episodes, featuring compelling true-crime mysteries, powerful documentaries and in-depth investigations. Follow now to get the latest episodes of Dateline NBC completely free, or subscribe to Dateline Premium for ad-free listening and exclusive bonus content: DatelinePremium.com

On Purpose with Jay Shetty

On Purpose with Jay Shetty

I’m Jay Shetty host of On Purpose the worlds #1 Mental Health podcast and I’m so grateful you found us. I started this podcast 5 years ago to invite you into conversations and workshops that are designed to help make you happier, healthier and more healed. I believe that when you (yes you) feel seen, heard and understood you’re able to deal with relationship struggles, work challenges and life’s ups and downs with more ease and grace. I interview experts, celebrities, thought leaders and athletes so that we can grow our mindset, build better habits and uncover a side of them we’ve never seen before. New episodes every Monday and Friday. Your support means the world to me and I don’t take it for granted — click the follow button and leave a review to help us spread the love with On Purpose. I can’t wait for you to listen to your first or 500th episode!

Music, radio and podcasts, all free. Listen online or download the iHeart App.

Connect

© 2025 iHeartMedia, Inc.