I’ve spent many hours over the past few weeks talking with my research clients on the phone. (I think this is an important part of what we do at Waverly Advisors–we don’t sell a “newsletter” or “service”; we have clients and those people are important to us, so we find the time, several times a year, to have conversations with them.) These calls are helpful for me because it gives me a little more insight into where people are focusing. Of course, some things were no surprise–people want to know about the Fed, impact of rate hikes, and the big question of whether this is a healthy pullback in stocks or the beginning of something bigger. What surprised me, however, was how often the question of stock market seasonality came up.

We deal with this question in the last quarter of the year every year. I’ve written some in-depth blogs in the past (a video here, here, and here, for instance), and I found myself having some conversations about significance testing (and the limitations of those tests.) Rather than dig into a math lesson here, I thought it might be productive to approach the problem from another perspective.

**Whenever anyone presents you with market stats, the key question is “how likely is this to affect the future?”** All other questions are really variations of that question. Sometimes people will throw out stats with arguments like “you can’t argue with math”, but it’s not really about the math, is it? It’s about the conclusions and implications we draw from the math, and *that* is certainly open to interpretation.

Patterns could exist for two reasons: 1) they point to something real in the data or 2) they simply arise due to random variation in the data. One of the best ways to build intuition about this concept is to create some simulations that are totally random–we know there is no actual influence and we know that each datapoint is basically a coin flip. Given those facts, let’s see what kind of patterns we can find in the data, and then compare those patterns to what we see in the real market data. If nothing else, it will give us new respect for how random data can appear to be non-random.

First, let’s look at the question of weekly seasonality in the S&P 500 index over the last 10 years. A good place to start is simply asking how many times a particular week (we could do the same thing with months, days, or any other time division) has been positive over those 10 years. Here’s one way to do it:

- First, we take weekly data for the S&P 500 for the past 10 years, and number the weeks 1-53. (There are 53 numbered weeks in some years.)
- Let’s then do a very simple analysis and score each week 1 if it was positive and 0 if not. (We are lumping unchanged weeks into down weeks, but that’s ok for simplicity.)
- For each week of the year, count the number that were up. (Call these “UP#”, just for reference.) We find, for instance, that week 1 was up 4/10 times, week 2 was positive 7/10, etc.
- Last, count how many UP#s fall into the bins 1-10. In other words, how many weeks were up only 1 year out of the last 10? How many were up 5/10?

Here’s what this data looks like for the last 10 years of the S&P 500. Very interesting–we have some weeks that have been up 8 out of the last 10 years, and one that was “almost always up”, declining only one year out of the past 10.

Is this “best week of the year” something we should focus on? Maybe we should just trade that week, since it was up an astounding 9 times out of the past 10 years, right? Maybe also those two weeks that were up eight times in 10 years–perhaps we should load up leveraged risk on those weeks and go nuts, right?

Well, let’s think a bit deeper. **There’s a chance that these results could be due to chance, and so a question we might ask is something like “how likely are we to see this kind of result if there is no actual seasonal influence?”** Another way to phrase the same thing would be to assume that markets are completely random, and then to see how often we’d find a week that was up 9 times in the past 10 years, and we can easily play with this in Excel.

I made a simulation that created random weeks (based on return assumptions drawn from actual market data, but you could also do the same thing with a coin flip), and then did the same type of analysis. Here are the first ten runs (columns R1 – R10) I got from that simulation (with the actual market data on the right side.)

Now, think carefully about what this simulation is and how it works. Each week is basically a coin flip–there is no way to predict that any week, on any subsequent year, is more likely to be up or down, because they are random. However, if we were to do “seasonal analysis”, we might be tempted to focus on the “really good” weeks. Some of the runs have weeks that were up all 10 years, and most have a handful that were up 8 or 9 times out of the past 10 years. Though these apparent seasonals are there in the data, they arose purely due to chance, and have no ability to tell us anything about the future. Does the real market data look any different?

There are better statistical tools (e.g., significance tests) to answer this question, but I think a simulation is a simple way to help us build some intuition. In this case, it’s easy to flip the lesson around and apply to it the “worst weeks of the year”–there’s simply no good statistical work that supports any non-random seasonal influences in the market. Yes, anyone can naively count weeks and produce stats, sell research and get reblogs based on these bad stats, but they do investors and traders a disservice. If there is no predictive power to the stats, then there’s no sense analyzing them. One of the worst things we can do is to incorporate bad information into our analysis, but that’s the topic of this week’s podcast.