# Solved – Variance of Random Matrix

Let's consider independent random vectors \$hat{boldsymboltheta}_i\$, \$i = 1, dots, m\$, which are all unbiased for \$boldsymboltheta\$ and that
\$\$mathbb{E}left[left(hat{boldsymboltheta}_i –
boldsymbolthetaright)^{T}left(hat{boldsymboltheta}_i –
boldsymbolthetaright)right] = sigma^2text{.}\$\$ Let
\$mathbf{1}_{n times p}\$ be the \$n times p\$ matrix of all ones.

Consider the problem of finding
\$\$mathbb{E}left[left(hat{boldsymboltheta} –
boldsymbolthetaright)^{T}left(hat{boldsymboltheta} –
boldsymbolthetaright)right]\$\$ where \$\$hat{boldsymboltheta} =
dfrac{1}{m}sum_{i=1}^{m}hat{boldsymboltheta}_itext{.}\$\$

My attempt is to notice the fact that \$\$hat{boldsymboltheta} = dfrac{1}{m}underbrace{begin{bmatrix}
hat{boldsymboltheta}_1 & hat{boldsymboltheta}_2 & cdots & hat{boldsymboltheta}_m
end{bmatrix}}_{mathbf{S}}mathbf{1}_{m times 1}\$\$
and thus
\$\$text{Var}(hat{boldsymboltheta}) = dfrac{1}{m^2}text{Var}(mathbf{S}mathbf{1}_{m times 1})text{.}\$\$
How does one find the variance of a random matrix times a constant vector? You may assume that I am familiar with finding variances of linear transformations of a random vector: i.e., if \$mathbf{x}\$ is a random vector, \$mathbf{b}\$ a vector of constants, and \$mathbf{A}\$ a matrix of constants, assuming all are comformable,
\$\$mathbb{E}[mathbf{A}mathbf{x}+mathbf{b}] = mathbf{A}mathbb{E}[mathbf{x}]+mathbf{b}\$\$
\$\$mathrm{Var}left(mathbf{A}mathbf{x}+mathbf{b}right)=mathbf{A}mathrm{Var}(mathbf{x})mathbf{A}^{prime}\$\$

Contents