2.4. Mean and variance propagation laws#

Here, will only consider the propagation of the mean (= expectation) and (co-)variances, and not the transformation of the full PDF or CDF.

The general form of our problem will be given as follows. Consider the single function of \(m\) random variables

\[ X = q(Y)=q(Y_1,\ldots,Y_m) \]

with the mean and covariance matrix of \(Y\) known:

\[ \mathbb{E}(Y)=\mu_Y, \quad \mathbb{D}(Y)=\Sigma_Y \]

Video#

In this video the first part of this chapter is explained. You can also first read this section, and only watch the video if you like someone to explain it.

What are the mean and variance of \(X\)?#

Let’s start with the mean for the case that the function is linear:

\[ q(Y)=a_1 Y_1+ a_2 Y_2 +\cdots+ a_m Y_m + c \]

with the \(a_i\) and \(c\) deterministic constants. Since the expectation operator is a linear operator, we have:

\[ \mathbb{E}(q(Y))=\mathbb{E}(a_1 Y_1+ a_2 Y_2 +\cdots a_m Y_m + c)= a_1 \mathbb{E}(Y_1)+\cdots+ a_m \mathbb{E}(Y_m)+c \]

But what if the function is non-linear? Then we can use the Taylor series approximation of \(q(Y)\).

Function of one random variable#

We will first look at the simplest case, where we have a function of a single random variable, \(X=q(Y)\), with the Taylor approximation:

\[ q(Y)\approx q(\mu_Y) +\left(\frac{\partial q}{\partial Y}\right)_0(Y-\mu_Y) + \frac{1}{2!} \left(\frac{\partial^2 q}{\partial Y^2}\right)_0(Y-\mu_Y)^2 + \text{H.O.T} \]

where we take the mean \(y_0=\mu_Y\) as the logical initial guess of the random vector \(Y\). (H.O.T. stands for higher-order terms). The subscript \(_0\) indicates that the partial derivatives are evaluated at \(y_0=\mu_Y\).

Due to the linearity of the expectation we then find as a second-order approximation of \(\mathbb{E}(q(Y))\), known as mean propagation law:

\[\begin{split} \begin{align*} \mathbb{E}(X)=\mathbb{E}(q(Y))&\approx \mathbb{E}(q(\mu_Y) +\left(\frac{\partial q}{\partial Y}\right)_0(Y-\mu_Y) + \frac{1}{2!} \left(\frac{\partial^2 q}{\partial Y^2}\right)_0(Y-\mu_Y)^2)\\ &= \mathbb{E}(q(\mu_Y)) +\left(\frac{\partial q}{\partial Y}\right)_0\mathbb{E}\left((Y-\mu_Y)\right) + \frac{1}{2} \left(\frac{\partial^2 q}{\partial Y^2}\right)_0\mathbb{E}\left((Y-\mu_Y)^2\right)\\ &= q(\mu_Y)+\frac{1}{2} \left(\frac{\partial^2 q}{\partial Y^2}\right)_0\sigma_Y^2 \end{align*} \end{split}\]

where in order to arrive at the last equation we should recognize that:

  • \(\mathbb{E}(q(\mu_Y))=q(\mu_Y)\) (since \(\mu_Y\) is deterministic and known)

  • \(\mathbb{E}(Y-\mu_Y)=\mathbb{E}(Y)-\mu_Y= 0\)

  • \(\mathbb{E}\left((Y-\mu_Y)^2\right)=\sigma_Y^2\).

For the variance of \(X=q(Y)\) it generally suffices to use a first-order approximation. The result follows as:

\[ \sigma_X^2 =\left(\frac{\partial q}{\partial Y}\right)_0^2\sigma_Y^2 \]

and is referred to as the variance propagation law.

Example \(X=Y^2\)
\[\begin{split} \begin{align*} \mathbb{E}(X)&\approx \mu_Y^2 + \frac{1}{2}\cdot 2 \cdot \sigma_Y^2= \mu_Y^2+\sigma_Y^2\\ \sigma_X^2 &\approx \left( 2\mu_Y\right)^2\sigma_Y^2 = 4\mu_Y^2\sigma_Y^2 \end{align*} \end{split}\]
Example gravitational force

Function of two random variables#

Let’s consider the case that we have one function of two random variables, \(Y = [Y_1\; \;Y_2]^T\) with known mean and covariance matrix:

\[\begin{split} \mathbb{E}(Y)=\mu_Y =\begin{bmatrix}\mu_1\\ \mu_2 \end{bmatrix}, \quad \Sigma_Y= \begin{bmatrix}\sigma_1^2 & Cov(Y_1,Y_2)\\ Cov(Y_1,Y_2)&\sigma_2^2 \end{bmatrix} \end{split}\]

The Taylor series approximations of \(X=q(Y_1,Y_2)\) follow as:

\[\begin{split} \begin{align*} \mathbb{E}(X)&\approx q(\mu_Y)+\frac{1}{2} \left(\frac{\partial^2 q}{\partial Y_1^2}\right)_0\sigma_1^2 +\frac{1}{2} \left(\frac{\partial^2 q}{\partial Y_2^2}\right)_0\sigma_2^2 + \left(\frac{\partial^2 q}{\partial Y_1 \partial Y_2}\right)_0Cov(Y_1,Y_2)\\ \sigma_X^2 &\approx \left(\frac{\partial q}{\partial Y_1}\right)_0^2\sigma_1^2 + \left(\frac{\partial q}{\partial Y_2}\right)_0^2\sigma_2^2 + 2\left(\frac{\partial q}{\partial Y_1}\right)_0\left(\frac{\partial q}{\partial Y_2}\right)_0Cov(Y_1,Y_2) \end{align*} \end{split}\]

These are thus the mean and variance propagation laws for a function of two random variables. Pay attention to the \(\approx\)-sign.

If \(Y_1\) and \(Y_2\) are independent, we have that \(Cov(Y_1,Y_2)=0\), such that the last term in both equations disappears.

Exercise mathematical pendulum

We will measure the period on one oscillation \(T\) of a pendulum, and also the length \(L\) of the pendulum. Both measurements are affected by random errors, and therefore the ‘calculated’ gravitational acceleration \(G\) is a function of two random variables:

\(G=q(L,T)= 4\pi^2 \frac{L}{T^2}\)

Apply the mean and variance propagation laws to approximate the mean and variance of \(G\) given that the expectations \(\mathbb{E}(L)= \mu_L\) and \(\mathbb{E}(T)= \mu_T\), as well as the variances \(\sigma^2_L\) and \(\sigma^2_T\) are known, and the covariance \(Cov(L,T)=0\).

Exercise moment of inertia

Functions of \(n\) random variables#

The propagation laws for functions of \(n\) random variables are as follows:

\[ \mathbb{E}(X)\approx q(\mu_Y)+\frac{1}{2} \sum_{i=1}^{n}\left(\frac{\partial^2 q}{\partial Y_i^2}\right)_0\sigma_i^2 + \frac{1}{2} \sum_{i=1}^{n}\sum_{j=1,j\neq i}^{n} \left(\frac{\partial^2 q}{\partial Y_i \partial Y_j}\right)_0Cov(Y_i,Y_j) \]
\[ \sigma_X^2 \approx \sum_{i=1}^{n}\left(\frac{\partial q}{\partial Y_i}\right)_0^2\sigma_i^2 + \sum_{i=1}^{n}\sum_{j=1,j\neq i}^{n}\left(\frac{\partial q}{\partial Y_i}\right)_0\left(\frac{\partial q}{\partial Y_j}\right)_0Cov(Y_i,Y_j) \]
Exercises wind load

We have the following equation, corresponding to the wind loading equation \(F\),

\[ F=A \cdot P \cdot C_d \]

Assume \(A\), \(P\), and \(C_d\) are independent random variables.

What is the covariance matrix of the random vector \(Y=[A, P, C_d]^T\)?

What is the approximation of the mean \(\mu_F\)?

What is the approximation of \(\sigma^2_F\)?

Exercise evaporation

The evaporation \(Q\) is given as function of net radiation \(R_n\), ground heat flux \(G\) and Bowen ratio \(B\):

\[ Q=\frac{R_n-G}{1-B} \]

Assume that the random variables \(R_n\), \(G\), \(B\) are independent.

What is the approximation of \(\sigma^2_Q\):

Video#

In this video the functions of \(n\) random variables is discussed, as well as the linear propagation laws is explained.

Attribution

This chapter was written by Sandra Verhagen. Find out more here.