Static Estimation
Static estimation estimates constant parameters that don’t change over time. In contrast, dynamic estimation estimates state variables that change over time.
Now, let’s examine what static estimation is.
Let there be two unknown vectors, a constant vector $X \in \mathbb{R}^{n}$ and a measurement vector $Z \in \mathbb{R}^{p}$. Since these two vectors are related, measuring $Z$ provides information about $X$. They can be expressed as a general non-linear equation, an additional measurement noise, and a linear equation as follows, respectively:
where $V$ is a random vector of sensor measurement noise, and $H \in \mathbb{R}^{p \times n}$. Since $V$ is a random vector, $Z$ is also a random vector.
For example, consider a simple system that estimates the 2D location of an object using radar. Assume that the object is fixed at $(x, y)$ and the radar can estimate the distance $R$ and azimuth $\Psi$. Then, $R$ and $\Psi$ are given as follows, respectively:
Since there are always measurement noises, the values of $R$ and $\Psi$ that the radar estimates are mixed with noises. Therefore, the measurement equations are:
where $V_{r}$ and $V_{\Psi}$ are the measurement noises of distance and azimuth respectively.
Expressing these formulas as vectors:
If the measurement is repeated several times, all the measurements are expressed as:
Where $k$ is time index.
Since measurements are taken independently, the measurement noises are independent and uncorrelated over time.
The combination of measurement vectors that collect all random measurement vectors from $k=0$ to $k$ is expressed as follows:
And the combination of the measured values of random measurement vectors are expressed as follows:
where
Static Estimation Problem
A static estimation problem is a problem that designs an estimator of constant vectors as a function of a combination of measurement vectors $z_{k}$ as follows:
If the measurement vectors are given as $Z_{k}=z_{k}$, the estimation values are deterministic; however, if the measurement vectors are given as random vectors, the estimation values are also random vectors as follows:
Depending on the unknown nature of $X$, the estimator is divided into two methods:
Bayesian approach
non-Bayesian approach
The evaluation elements of an estimator are:
Bias
Consistency
Covariance of estimation errors
Bayesian Approach
In a Bayesian approach, $X$ is a random vector. Therefore, assume that we know a priori probability information of $X$.
A measurement vector $Z$ reinforces the probability information of $X$ more precisely.
MAP estimator and MMSE estimator are a Bayesian approach estimators. The basic form of Bayesian approach estimator is :
non-Bayesian Approach
In a non-Bayesian approach, $X$ is an unknown deterministic value. Therefore, assume that there is no a priori probability information, and the information of $X$ can only be obtained by a measurement vector $Z$.
ML estimator and WLS estimator are non-Bayesian approach estimators. The basic form of non-Bayesian approach estimators is:
Bias
When an unknown constant vector $X$ is a random vector and probability information is given as a probability density function $p_{X}(x)$, if there is an estimator such that $\mathbb{E}[\hat{X}] = \mathbb{E}[X]$, it is called an unbaised estimator. Alternatively, if the estimator is $\mathbb{E}[\hat{X}] = x$ where $X$ is an unknown constant vector which is $x$, it is also called an unbaised estimator.
When $k \rightarrow \infty$, if $\mathbb{E}[\hat{X}] = \mathbb{E}[X]$ or $\mathbb{E}[\hat{X}] = x$ are established, it is called an asymptotically unbiased estimator.
Consistency
When an unknown constant vector $X$ is a random vector, if there is an estimator such that $\lim_{k \rightarrow \infty}\mathbb{E}[(X-\hat{X}(k))^{T}(X-\hat{X}(k))] = 0$, it is called a consistent estimator. Alternatively, if the estimator is $\lim_{k \rightarrow \infty}\mathbb{E}[(x-\hat{X}(k))^{T}(x-\hat{X}(k))] = 0$ where $X$ is an unknown constant vector which is $x$, it is also called a consistent estimator.
Covariance of Estimation Errors
Covariance of estimation errors is used to evaluate the quality of an estimator. That is, the smaller the covariance value, the better the estimate.
When an unknown constant vector $X$ is a random vector, the covariance of estimation errors of an unbiased estimator is given as:
When $X$ is an unknown constant vector which is $x$, the covariance of estimation errors of an unbiased estimator is given as: