Consider exchangeable random variables . A couple of facts seem quite intuitive:

**Statement 1.** The “variability” of sample mean decreases with .

**Statement 2.** Let the average of functions be defined as . Then is less “variable” than .

To make these statements precise, one faces the fundamental question of comparing two random variables and (or more precisely comparing two distributions). One common way we think of ordering random variables is the notion of stochastic dominance:

However, this notion really is only a suitable notion when one is concerned with the actual size of the random quantities of interest, while, in our scenario of interest, a more natural order would be that which compares the variability between two random variables (or more precisely, again, the two distributions). It turns out that a very useful notion, used in a variety of fields, is due to Ross (1983): Random variable is said to be *stochastically less variable* than random variable (denoted by ) when every risk-averse decision maker will choose over (given they have similar means). More precisely, for random variables and with finite means

where is the set of functions for which the above expectations exist.

One interesting, but perhaps not entirely obvious, fact is that this notion of ordering is equivalent to saying that there is a sequence of mean preserving spreads that in the limit transforms the distribution of into the distribution of another random variable with finite mean such that ! Also, using results by Hardy, Littlewood and Polya (1929), the stochastic variability order introduced above can be shown to be equivalent to Lorenz (1905) ordering used in economics to measure income equality.

Now with this, we are ready to formalize our previous statements. The first statement is actually due to Arnold and Villasenor (1986):

Note that when you apply this fact to a sequence of iid random variables with finite mean , it strengthens the strong law of large number in that it ensures that the almost sure convergence of the sample mean to the mean value occurs with monotonically decreasing variability (as the sample size grows).

The second statement comes up in proving certain optimality result in sharing parallel servers in fork-join queueing systems (J. 2008) and has a similar flavor:

The cleanest way to prove both statements, to the best of my knowledge, is based on the following theorem first proved by Blackwell in 1953 (later strengthened to random elements in separable Banach spaces by Strassen in 1965, hence referred to by some as Strassen’s theorem):

**Theorem 1** * Let and be two random variables with finite means. A necessary and sufficient condition for is that there are two random variables and with the same marginals as and , respectively, such that almost surely. *

For instance, to prove the first statement we consider and . All that is necessary now is to note that , is an independent uniform rv on the set , has the same distribution as random variable . Furthermore,

Similarly to prove the second statement, one can construct by selecting a random permutation of functions .