Categories
Research Methodologies
April 22, 2020
A rebuttal to the call to return to simple linear data models.
Some days you read an article or a blog post that makes you shake your head – you groan and go on to the next thing. Then there are the other days when you read something that makes you want to scream, “NO NO NO!”. That’s what happened when I read the recent call for us to return to “neo-classical” models to solve our business problems. Larry’s introduction got me excited, as it echoed my own sentiments about our industry’s boundless enthusiasm for things new and shiny, regardless of how well they work. Alas, the intro was the best part of this blog.
The author claims that a lot of the newer, more complex modeling techniques don’t work well and may not be an improvement over simple linear models. He begins to explain why he believes this and I’m ready to bite – I’ve seen lots of models fail for any number of reasons and I’m hoping he has some new ammunition for me. I’m even ready to accept a re-hash of the topic. Maybe he’ll point out that these complex models can be black-boxy or they are often over-specified or they are inappropriate for the data source. Instead, I get six reasons why we should go back to simple linear models – not a single one of which is legitimate. Let’s look at the reasons the author brings up:
The only reason you adopt a simple linear model is that the relationships in your data are reasonably linear. If the relationships between variables are non-linear, then you don’t use a linear model. Let me give you an example. If you look at the relationship between sales and the price spread of a private label product against its national brand, you’ll find an inverted U-shaped function. There’s a sweet spot for the price gap at the top of the curve; moving away from that optimal price gap, either by increasing or decreasing the price spread, hurts sales. If you use a simple linear model, you get a non-significant result because the correlation between price spread and sales is zero. You need a more complex, non-linear model to describe this data.
John Tukey wrote his book Exploratory Data Analysis back in 1977 – to encourage us to look at the shape of our data before analyzing it. I write this because this blog is filled with really bad advice and can only lead you, the researcher, down a dangerous road (even if you had good intentions, you know where that road takes you).
A good researcher will read the above, shake her head, and move on. A smart researcher will ask an expert about the criteria for good models. A really smart researcher will look up Kevin Gray. He writes often and well about what makes for good and bad models in a language that you don’t need a Stat degree to comprehend.
Comments
Comments are moderated to ensure respect towards the author and to prevent spam or self-promotion. Your comment may be edited, rejected, or approved based on these criteria. By commenting, you accept these terms and take responsibility for your contributions.
Disclaimer
The views, opinions, data, and methodologies expressed above are those of the contributor(s) and do not necessarily reflect or represent the official policies, positions, or beliefs of Greenbook.
More from Steve Needel,
Since the start of COVID, marketing research blogs have been forecasting dire consequences for the industry. Enough already.
An argument for the importance of quality over quantity in data.
A rebuttal by Steve Needel defending the validity of significance testing.
Sign Up for
Updates
Get content that matters, written by top insights industry experts, delivered right to your inbox.
67k+ subscribers