A.21  Estimation Theory

A.21.1  Probabilistic estimation theory

Estimation theory concerns estimating the values of parameters 𝜃 based on measured data. An estimator attempts to approximate the unknown parameters using measurements that have a random component (e. g., due to noise). There are other approaches, but the one of interest to this text is the probabilistic estimation theory, which aims at finding a set Θ = {𝜃1,,𝜃M} of M parameters. In this approach, the measured data set Ξ with N elements is random and with a probability distribution p(Ξ|Θ) that depends on Θ. The parameters 𝜃 themselves may have associated probability distributions. An estimator provides Θ^ = {𝜃^i}, where the “hat” suggests the element 𝜃^i is an estimate of the true parameter 𝜃i. The value 𝜃^i is a statistics with distribution, variance, mean 𝔼[𝜃^i], etc.

The bias of an estimator is 𝔼[𝜃^i] 𝜃i and an estimator is called unbiased if the bias is zero for all M elements of Θ^. The estimator of the sample variance that uses N as normalization factor instead of N 1 is a good example of a biased estimator [ urlBMbia]. There is a well-known tradeoff between bias and variance. In many applications it is possible to decrease the variance of an estimator at the expenses of allowing its bias to increase.

Some specific topics of probabilistic estimation theory are discussed in the sequel. But this theory is broad and some of the (many) concepts out of the scope of this text are:

A.21.2  Minimum mean square error (MMSE) estimators

A popular estimator is the minimum mean square error (MMSE) estimator, which utilizes the error ei = 𝜃i 𝜃^i,i = 1,,M, between the estimated parameters and their actual values.

The measurements Ξ may be organized in several different forms. For example, in multivariate statistics, each element of Ξ can be an array. For simplicity, it is assumed here that Ξ and Θ are column vectors y and x with N and M complex-valued elements, respectively. The estimator x^(y) is a function of the measurements y and the estimation error is e = x^ x with elements ei,i = 1,,M. Because the vectors are random, the MSE is given by

MSE=def i=1M𝔼[|e i|2],
(A.95)

which can be written in matrix notation as the expected value of the trace

MSE = tr {𝔼 [eeH]} ,
(A.96)

where Ree = 𝔼[eeH] is the error autocorrelation matrix.

A.21.3  Orthogonality principle

AK-TODO Orthogonality principle