A.21 Estimation Theory
A.21.1 Probabilistic estimation theory
Estimation theory concerns estimating the values of parameters based on measured data. An estimator attempts to approximate the unknown parameters using measurements that have a random component (e. g., due to noise). There are other approaches, but the one of interest to this text is the probabilistic estimation theory, which aims at finding a set of parameters. In this approach, the measured data set with elements is random and with a probability distribution that depends on . The parameters themselves may have associated probability distributions. An estimator provides , where the “hat” suggests the element is an estimate of the true parameter . The value is a statistics with distribution, variance, mean , etc.
The bias of an estimator is and an estimator is called unbiased if the bias is zero for all elements of . The estimator of the sample variance that uses as normalization factor instead of is a good example of a biased estimator [ urlBMbia]. There is a well-known tradeoff between bias and variance. In many applications it is possible to decrease the variance of an estimator at the expenses of allowing its bias to increase.
Some specific topics of probabilistic estimation theory are discussed in the sequel. But this theory is broad and some of the (many) concepts out of the scope of this text are:
- Cramér-Rao bound (CRB): lower bound on the variance of an estimator;
- Several estimators: Maximum likelihood (MLE), maximum a posteriori (MAP), generalized Bayes, etc.;
- Non-linear estimators such as artificial neural networks;
- Dynamic systems: Kalman filter, recursive Bayesian estimation, etc.
A.21.2 Minimum mean square error (MMSE) estimators
A popular estimator is the minimum mean square error (MMSE) estimator, which utilizes the error between the estimated parameters and their actual values.
The measurements may be organized in several different forms. For example, in multivariate statistics, each element of can be an array. For simplicity, it is assumed here that and are column vectors and with and complex-valued elements, respectively. The estimator is a function of the measurements and the estimation error is with elements . Because the vectors are random, the MSE is given by
| (A.95) |
which can be written in matrix notation as the expected value of the trace
| (A.96) |
where is the error autocorrelation matrix.
A.21.3 Orthogonality principle
AK-TODO Orthogonality principle