Bayesian filters make use of Bayesian Probability and prior knowledge to obtain a probability distribution function (PDF) that can be used to infer its state amidst limited information
Relationship between Bayesian and Kalman filters
Kalman filters belong to the family of Bayesian filters. They are basically recursive Bayesian filters which assume that all distributions are Gaussian and apply to linear Gaussian state space models
The working of a Bayesian filter is summarized in this diagram [citation needed]
\begin{flalign*} \texttt{Prediction: } \bar{\textbf{x}} = \textbf{x} * f_x(\bullet) \\ \texttt{Update: } ||{\mathcal{L}}\cdot{x}|| \end{flalign*}- The initialization step, as the name says, is where our belief in the state is initialized
- The prediction step follows next. Here, the model of the system is used to estimate the value for the next system state. Any uncertainties with the belief will be accounted for here. The prediction is done by performing a convolution of a kernel function over the PDF
- The next step is the update step. A measurement and its associated belief is taken and the likelihood is calculated based on that. the posterior is calculated using the Bayes formula
One cycle of prediction and updating with a measurement is called a system evolution
Looping the prediction step
If the sensor is noisy, we will always lose some information in the prediction. If we keep looping the prediction step on this, we will eventually end up with no information left, and the probabilities will be evenly distributed across the distribution
For most tracking and filtering applications, we want a unimodal and continuous filter. Multimodal filtering is computationally expensive and counterintuitive