A.23 Vector prediction exploring spatial correlation
Instead of exploring correlation over time, this section discusses methods to explore the so-called spatial correlation: an element of the random vector being estimated based on the other elements of this vector. As signal model, the Gaussian block or “packet” ISI channel [CF97], (page 84, Eq. 4.6) is adopted here, which is given by
| (A.103) |
where is a zero-mean complex random input -vector, is a zero-mean complex random output -vector, is a complex zero-mean Gaussian noise -vector independent of and is the complex channel matrix [CF97]. For these random vectors, the correlation matrices coincide with covariance matrices. 20
The output covariance matrix is given by
| (A.104) |
where a matrix superscript denotes Hermitian (transpose conjugate).
A characteristic representation of a random -vector is given by the linear combination of the columns of a matriz , whose determinant is equal to one,21 weighted by a vector of uncorrelated random variables , that is:
| (A.105) |
Hence, the covariance matrix of is given by:
| (A.106) |
where is diagonal (the random variables in are uncorrelated).
There are two alternatives of interest for representing a vector in its characteristic form, the modal and the innovations representation. The first is derived from the eigendecomposition of . Given the factorization , by comparison to Eq. (A.106), corresponds to the unitary matrix from the eigendecomposition, while the uncorrelated vector from Eq. (A.105) corresponds to . Meanwhile, the latter (innovations representation) is derived from the Cholesky decompostion. In a similar manner, given the factorization , corresponds to the lower triangular matrix , while corresponds to .
The important conclusion yielded by these two representations is that a vector whose random variables are correlated can be whitened by a forward section given by , the inverse of the unitary matrix from the eigendecomposition of its covariance matrix, or , the inverse of the lower triangular matrix from the Cholesky decomposition of its covariance matrix.
The innovations representation is a natural adaptation of linear prediction over time and is obtained with a Cholesky factorization of , while the modal representation can be obtained via eigenanalysis or SVD.
The optimum MMSE linear predictor in this scenario is
| (A.107) |
where the predictor matrix is given by
| (A.108) |
with being obtained from the innovations representation and being the identity matrix. It was assumed that is nonsingular, otherwise the pseudo inverse can be used.
Because is lower triangular and monic, its inverse is also lower triangular and monic. The subtraction of from makes to be lower triangular with zeros in the main diagonal. This structure imposes a causal relation among the elements of , such that can be obtained recursively.
The error vector is
| (A.109) |
In general, the sum mean-squared prediction error is
| (A.110) |
where is the autocorrelation matrix of . It can be proved (see, e. g., [BLM04]) that when the optimum linear predictor of Eq. (A.108) is adopted, the error power achieves its minimum value given by . This avoids the step of estimating to obtain the prediction gain, which is given by
| (A.111) |
Hence, making an analogy with prediction over time, repeated here for convenience:
the spatial prediction allows to obtain
which is expressed in matrix notation as and .
Listing A.15 illustrates an example discussed in [BLM04].
1%Example 10-11 from Barry, 2004 (note a typo in matrix R in the book) 2Ryy=[16 8 4; 8 20 10; 4 10 21] %noise autocorrelation, correlated 3%Ryy=[10, 8, 2; 8 10 10; 2 10 10]; %another option, higher gain 4[L D] = ldl_dg(Ryy)%own LDL, do not use chol(A) because it swaps rows 5Ryy-L*D*L' %compare with Ryy, should be the same 6P=eye(size(L))-inv(L) %optimum MMSE linear predictor 7minMSE=trace(D) %minimum MSE is the trace{Ree} = trace{D} 8sumPowerX=trace(Ryy); %sum of all "users" 9predictionGain = 10*log10(sumPowerX/minMSE)
In Listing A.15, the original predictor matrix is
P = [0, 0, 0; 0.5000, 0, 0; 0, 0.5000, 0]
and the prediction gain is 0.7463 dB. Adopting a new correlation matrix Ryy=[10, 8, 2; 8 10 10; 2 10 10] leads to
P = [0, 0, 0; 0.8000, 0, 0; -1.6667, 2.3333, 0]
and a prediction gain of 9.2082 dB.
Note that the first element of in is always zero due to the structure of . Then, the second element is a scaled version of the first element of , and so on.