Variance of Discrete Random Variable from PGF

Theorem
Let $X$ be a discrete random variable whose probability generating function is $\map {\Pi_X} s$.

Then the variance of $X$ can be obtained from the second derivative of $\map {\Pi_X} s$ $s$ at $x = 1$:


 * $\var X = \map {\Pi''_X} 1 + \mu - \mu^2$

where $\mu = \expect X$ is the expectation of $X$.

Proof
From the definition of the probability generating function:
 * $\ds \map {\Pi_X} s = \sum_{x \mathop \ge 2} \map p x s^x$

From Derivatives of Probability Generating Function at One:
 * $\ds \map {\Pi''_X} s = \sum_{x \mathop \ge 2} x \paren {x - 1} \map p x s^{x - 2}$

But it also holds when you include $x = 0$ and $x = 1$ in the sum, as in both cases the term evaluates to zero and therefore vanishes.

So:
 * $\ds \map {\Pi''_X} s = \sum_{x \mathop \ge 0} x \paren {x - 1}\map p x s^{x - 2}$

Plugging in $s = 1$ gives:

The result follows from the definition of variance:
 * $\var X = \expect {X^2} - \paren {\expect X}^2$

after a little algebra.

Comment
So, in order to find the variance of a discrete random variable, then there is no need to go through the tedious process of what might be a complicated and fiddly summation.

All you need to do is differentiate its PGF twice, and plug in $1$.

Assuming, of course, you know what its PGF is.