Suppose one day, I say to you that there’s a 20% chance of rain next weekend. Or that there’s a 20% chance of Donald Trump being elected for another term as President of the United States. Upon reading about some scandal the next day, I revise my prediction to 85%. By nighttime the matter settles, so I announce my updated prediction of 5%. In the end, it doesn’t rain (or Joe Biden defeats the incumbent).

After following a bunch of my forecasts, you might criticize me for being too fickle, my predictions swinging wildly as if I can’t make up my mind. Or you might think the opposite: that I play it too safe, only shifting my opinion to one side when the evidence becomes overwhelming.

What’s the right amount of variation? Is it a thing that can be measured?

Yes it is! In this blog post, I propose the sum of squared changes as such a measure.

First, a formal derivation. If you’re not into the technicalities of probability theory, the next section can be skipped. We’ll understand its implications afterward.

# A formal interlude

Let $$X_t$$ be a bounded martingale adapted to the filtration $$\mathcal F_t$$,1 and let $$0 = T_0 \le T_1 \le \ldots \le T_N$$ be stopping times. Then,

\begin{aligned} \mathbb E\Big( \sum_{n=1}^N (X_{T_n} - X_{T_{n-1}})^2 \mid \mathcal F_0\Big) &= \mathbb E\Big(X_{T_N}^2 - X_{0}^2 + 2 \sum_{n=1}^N X_{T_{n-1}} (X_{T_{n-1}} - X_{T_n}) \mid \mathcal F_0\Big) \\&= \mathbb E(X_{T_N}^2 \mid \mathcal F_0) - X_{0}^2 + 2 \sum_{n=1}^N \mathbb E\Big(X_{T_{n-1}} (X_{T_{n-1}} - X_{T_n}) \mid \mathcal F_0\Big) \\&= \mathbb E(X_{T_N}^2 \mid \mathcal F_0) - X_{0}^2 + 2 \sum_{n=1}^N \mathbb E\Big(X_{T_{n-1}} \underbrace{\mathbb E(X_{T_{n-1}} - X_{T_n} \mid \mathcal F_{T_{n-1}})}_{=0} \mid \mathcal F_0\Big) \\&= \mathbb E(X_{T_N}^2 \mid \mathcal F_0) - X_{0}^2 \end{aligned}

Let $$X_t$$ be the $$\mathcal F_t$$-conditioned probability of some event whose outcome is determined by time $$T_N$$. This is a martingale for which $$X_{T_N}$$ is $$1$$ with probability $$X_0$$, and $$0$$ otherwise. Hence, $$\mathbb E(X_{T_N}^2 \mid \mathcal F_0) = X_0$$.

# And we’re back!

So what have we shown? In essence, the martingale $$X_t$$ represents the prediction at time $$t$$. The role of the stopping times is to grant the forecaster some flexibility: they don’t have to announce predictions at every time $$t$$, nor even at pre-specified intervals. All that matters is that the decision on whether or not to announce a prediction at time $$t$$ be made by time $$t$$; deciding to withhold a prediction based on information from the future would be cheating! As long as this rule is satisfied, we can just add up the squared changes in the forecaster’s predictions to get a measure of their fickleness. If the forecaster reports genuine conditional probabilities, the first of which was $$X_0$$, then on average this sum should equal $$X_0 - X_0^2$$.

Using the example we started with, $$X_0 = 0.2$$, so $$X_0 - X_0^2 = 0.16$$. Our prediction series was $$(0.2,\, 0.85,\, 0.05,\, 0)$$, so the sum of squared changes is

$(0.2 - 0.85)^2 + (0.85 - 0.05)^2 + (0.05 - 0)^2 = 1.065$

This is a lot more than the expected $$0.16$$. Of course, since this is a random quantity, a lot more data would be needed to support the fickleness critique. The issue of statistical significance is beyond the scope of this blog post, but hopefully this discussion serves as a helpful start. As an exercise, you can try measuring this quantity for the prediction series of a well-known forecaster, such as FiveThirtyEight!

Now, let’s say you don’t know which event will come to pass, but you’re confident that the prediction market is more volatile than a true martingale, meaning that its sum of squared deviations will exceed the expected $$X_0 - X_0^2$$. Is it possible to bet on this outcome, making a profit if it comes to pass?
1. Filtration is a very technical term, but you can think of $$\mathcal F_t$$ as all of the information known at time $$t$$.