This is the lecture notes for "ELEC 5650: Networked Sensing, Estimation and Control" in the 2024-25 Spring semester, delivered by Prof. Ling Shi at HKUST. In this session, we will explore maximum a posteriori (MAP) estimation, minimum mean squared error (MMSE) estimation, maximum likelihood (ML) estimation, weighted least squares estimation, and linear minimum mean square error (LMMSE) estimation.
MAP (Maximum A Posterior) Estimation
\(x\) is the parameter to be estimated
\[
\hat x=\arg\max_x \begin{cases}
f(x|y),&x\text{ is continuous} \\
p(x|y),&x\text{ is discrete} \\
\end{cases}
\]
MMSE (Minimum Mean Squared Error) Estimation
\[
\hat x=\arg\min_{\hat x}E[e^Te|y]=\arg\min_{\hat x}E[\hat x|y],\ e=x-\hat x
\]
Non Bayesian. \(p(y|x)\) is conditional probability and \(p(y;x)\) is parameterized probability, \(p(y|x)\not\Leftrightarrow p(y;x)\).
Assume we have \(n\) measurements \(\mathcal X=(X_1,\cdots,X_n)\), we use \(p(\mathcal X;\theta)\) to describe the joint probability of \(\mathcal X\).
\[
\hat\theta_n=\arg\max_\theta\begin{cases}
f(\mathcal X;\theta),&\theta\text{ is continuous} \\
p(\mathcal X;\theta),&\theta\text{ is discrete} \\
\end{cases}
\]
This shows that error \(e=x-\hat x\) is independent of observation \(y\).
Innovation Process
Calculating \(\Sigma_{yy}\) consumes lots of time, however, if \(\Sigma_{yy}\) is diagonal the thing becomes easy. By G.S. process, we can obtain orthogonality vectors \(\vec e_1,\cdots\vec e_k\) and the lower triangular transform matrix \(F\) from \(\vec y_1,\cdots,\vec y_k\). The key idea of orthogonal projection is to decompose the observation vector \(y_k\) into a part related to the past prediction value, which can be predicted by \(y_1,\cdots y_{k-1}\), and a new part that is irrelevant to the past prediction value (innovation).