![]() |
(1) |
Our goal is to derive the form of the matrix, A, so that the expected mean square difference between the estimated field and the actual field ( x ) is minimized,
![]() |
(2) |
If we put (1) into (2) and expand, we get
![]() |
(3) |
If we let Cx be the autocorrelation of the field ( E[x xT] ),
be the autocorrelation of the observations
(
), and
be the
cross correlation between the field and the observations
(
), then we can write the above as
![]() |
(4) |
The next step requires the application of the following matrix identity (proved in the appendix),
(A - B C-1) C (A - B C-1)T - B C-1 BT = A C AT - B AT - (B AT)T | (5) |
using A in (4) for A in (5), and
for B as
well as
for C, we can reduce (x) to
![]() |
(6) |
The matrices ,
is an autocorrelation matrix therefore
both it and
are nonnegative definite (see
appendix), therefore
![]() |
(7) |
![]() |
(8) |
![]() |
(9) |
![]() |
(10) |
![]() |
(11) |
Further, we can write down what the expected error for the estimator
as,
![]() |
(12) |
Equations (11) and (12) constitute the Gauss-Markov estimator for the linear minimum means square estimate of a random variable.