The MMSE estimator is the mean of the posterior pdf \(E(x|y)\) of \(x\) given observation \(y\).
- The estimator is unbiased.
- The covariance is reduced compared to the a priori information.
- Commutes over affine transformation.
- Additivity property for independent data sets.
- Linear in the Gaussian case.
- The estimator error is orthogonal to the space spanned by all Y-measurable functions (affine functions being a subset)
The MAP estimator \( \textsf{arg max}_\theta \; p(\theta|x) \) given observation \(x\)
- Jointly Gaussian case, MAP = MMSE (posterior is Gaussian, hence pdf unimodal and symmetric, mean = mode = median)
- Do not commute over nonlinear transformation. (Invariant property does not hold, unlike ML)
- Commutes over linear transformation.
MAP tends to ML when
- Prior is uninformative
- Large amount of information in data compared to prior
Gaussian linear model
Let the observed samples takes on the model
\[ x = H\theta + w\] with prior \(\mathcal{N}(\mu_\theta, C_\theta)\) and noise vector \(\mathcal{N}(0, C_w)\) independent of \(\theta\), then the posterior is Gaussian with mean
\[ E(\theta|x) = \mu_\theta + C_\theta H^T (H C_\theta H^T + C_w)^{-1} (x - H \mu_\theta) \] and covariance \[ C_{\theta|x} = C_\theta - C_\theta H^T (H C_\theta H^T + C_w)^{-1} H C_\theta \] Contrary to the classical Gaussian linear model \(H\) does not need to be full rank.
In alternative form,
\[ E(\theta|x) = \mu_\theta + ( C_\theta^{-1} + H^T C_w^{-1} H )^{-1} H^T C_w^{-1} (x - H \mu_\theta)\] and \[ C_{\theta|x} = ( C_\theta^{-1} + H^T C_w^{-1} H )^{-1} \]
LMMSE estimator \( E^*[X|Y] \)
- A function of first and second order statistics only. \[E^*[X|Y] = \mu_x + \Sigma_{xy} \Sigma_{yy}^{-1} ( y - \mu_y) \] (inverse can be replaced with pseudo-inverse if necessary)
- Jointly Gaussian case, \(E^*[X|Y] = E[X|Y]\)
- Error orthogonal to subspace spanned by \(Y\)
- Additivity property \[E^*[X|Y_1,\dotsc,Y_k] = \sum_{j=1}^k E^*[X|Y_j] - (k-1)\mu_x \]
No comments:
Post a Comment