tex v2 - without experiments
This commit is contained in:
@@ -1,22 +1,17 @@
|
||||
\section{IMMPF and Mixing}
|
||||
\label{sec:immpf}
|
||||
|
||||
In the previous section we have introduced a standard particle filter, an evaluation step and two different transition models.
|
||||
Using this, we are able to implement two different localisation schemes.
|
||||
One providing a high diversity with a robust, but uncertain position estimation.
|
||||
The other keeps the localisation error low by using a very realistic propagation model, while being prone to sample impoverishment \cite{Ebner-15}.
|
||||
In the following, we will combine those filters using the Interacting Multiple Model Particle Filter (IMMPF) and a non-trivial Markov switching process.
|
||||
|
||||
%Einführen des IMMPF
|
||||
Consider a jump Markov non-linear system that is represented by different particle filters as state space description, where the characteristics change in time according to a Markov chain.
|
||||
The posterior distribution is then described by
|
||||
%
|
||||
\begin{equation}
|
||||
p(\mStateVec_{t}, m_t \mid \mObsVec_{1:t}) = P(m_k \mid \mObsVec_{1:t}) p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t})
|
||||
p(\mStateVec_{t}, m_t \mid \mObsVec_{1:t}) = P(m_t \mid \mObsVec_{1:t}) p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t})
|
||||
\label{equ:immpfPosterior}
|
||||
\end{equation}
|
||||
%
|
||||
where $m_t\in M\subset \mathbb{N}$ is the modal state of the system \cite{Driessen2005}.
|
||||
where $m_t\in M\subset \mathbb{N}$ is the modal state of the system and thus describes the current mode (particle filter) \cite{Driessen2005}.
|
||||
The notation $P(\cdot)$ provides a discrete probability distribution.
|
||||
Given \eqref{equ:immpfPosterior} and \eqref{equ:bayesInt}, the mode conditioned filtering stage can be written as
|
||||
%
|
||||
\begin{equation}
|
||||
@@ -31,7 +26,7 @@ Given \eqref{equ:immpfPosterior} and \eqref{equ:bayesInt}, the mode conditioned
|
||||
\label{equ:immpfFiltering}
|
||||
\end{equation}
|
||||
%
|
||||
and the posterior mode probabilities are calculated by
|
||||
and the posterior mode probabilities, providing how likely it is that a considered mode represents the system's state, are calculated by
|
||||
%
|
||||
\begin{equation}
|
||||
p(m_t \mid \mObsVec_{1:t}) \propto p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t-1}) P(m_t \mid \mObsVec_{1:t-1})
|
||||
@@ -40,7 +35,7 @@ and the posterior mode probabilities are calculated by
|
||||
\end{equation}
|
||||
%
|
||||
It should be noted that \eqref{equ:immpfFiltering} and \eqref{equ:immpModeProb} are not normalized and thus such a step is required.
|
||||
To provide a solution for $P(m_t \mid \mObsVec_{1:t-1})$, the recursion for $m_t$ in \eqref{equ:immpfPosterior} is now derived by the mixing stage \cite{Driessen2005}.
|
||||
To provide a solution for the probability distribution $P(m_t \mid \mObsVec_{1:t-1})$, the recursion for $m_t$ in \eqref{equ:immpfPosterior} is now derived by the so called mixing stage \cite{Driessen2005}.
|
||||
Here, we compute
|
||||
%
|
||||
\begin{equation}
|
||||
@@ -68,14 +63,26 @@ and
|
||||
\end{equation}
|
||||
%
|
||||
where \eqref{equ:immpModeMixing} is a weighted sum of distributions and the weights are provided through \eqref{equ:immpModeMixing2}.
|
||||
The transition probability $P(m_{t+1} = k \mid m_t = l)$ is given by the Markov transition matrix $[\Pi_t]_{kl}$.
|
||||
Sampling from \eqref{equ:immpModeMixing} is done by first drawing a modal state $m_t$ from $P(m_t \mid m_{t+1}, \mObsVec_{1:t})$ and then drawing a state $\mStateVec_{t}$ from $p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t})$ in dependence to that $m_t$.
|
||||
To find a solution for $P(m_t \mid \mObsVec_{1:t})$, an estimate of the posterior probability $p(m_t \mid \mObsVec_{1:t})$ in \eqref{equ:immpModeProb} can be calculated according to
|
||||
%
|
||||
\begin{equation}
|
||||
P(m_t \mid \mObsVec_{1:t}) = \frac{\omega_t^{m_t} P(m_t \mid \mObsVec_{1:t-1})}{\sum_{m=1}^M \omega_t^{m_t} P(m_t \mid \mObsVec_{1:t-1})}
|
||||
\enspace .
|
||||
\label{equ:immpMode2}
|
||||
\end{equation}
|
||||
%
|
||||
Here, $\omega_t^{m_t}$ is the unnormalized weight given by the state evaluation of the respective mode $m_t$.
|
||||
The initial mode probabilities $P(m_1 \mid \mObsVec_{1})$ have to be defined beforehand.
|
||||
The transition probability $P(m_{t+1} = k \mid m_t = l)$ in \eqref{equ:immpModeMixing3} is given by the Markov transition matrix $[\Pi_t]_{kl}$.
|
||||
The matrix $\Pi_t$ is a real square matrix, with each row summing to 1.
|
||||
It provides the probability of moving from $m_t$ to $m_{t+1}$ in one time step.
|
||||
Sampling from \eqref{equ:immpModeMixing} is done by first drawing a modal state $m_t$ from $P(m_t \mid m_{t+1}, \mObsVec_{1:t})$ and then drawing a state $\mStateVec_{t}$ from $p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t})$ in regard to that $m_t$.
|
||||
In context of particle filtering, this means that \eqref{equ:immpModeMixing} enables us to pick particles from all available modes in regard to the discrete distribution $P(m_t \mid m_{t+1}, \mObsVec_{1:t})$.
|
||||
Further, the number of particles in each mode can be selected independently of the actual mode probabilities.
|
||||
|
||||
Algorithm \ref{alg:immpf} shows the complete IMMPF procedure in detail.
|
||||
As prior knowledge, $M$ initial probabilities $P(m_1 \mid \mObsVec_{1})$ and initial distributions $p(\mStateVec_{1} \mid m_1, \mObsVec_{1})$ providing a particle set $\{W^i_{1}, \vec{X}^i_{1} \}_{i=1}^N$ are available.
|
||||
The mixing step requires that the independent running filtering process are all finished.
|
||||
The mixing step requires that the independently running filtering processes are all finished.
|
||||
|
||||
\begin{algorithm}[t]
|
||||
\caption{IMMPF Algorithm}
|
||||
@@ -106,31 +113,32 @@ The mixing step requires that the independent running filtering process are all
|
||||
|
||||
|
||||
%grundidee warum die matrix so gewählt wird.
|
||||
With the above, we are finally able to combine the two filters described in section \ref{sec:rse}.
|
||||
The basic idea of our approach is to utilize the restrictive filter as the dominant one, providing the state estimation for the localisation.
|
||||
Due to its robustness and good diversity the other, more permissive one, is then used as support for possible sample impoverishment.
|
||||
If we recognize that the dominant filter gets stuck or loses track, particles from the supporting filter will be picked with a higher probability while mixing the new particle set for the dominant filter.
|
||||
With the above, we are finally able to combine the two filters described in section \ref{sec:rse} and realize the considerations made in section \ref{sec:divergence}.
|
||||
Within the IMMPF we utilize the restrictive graph-based filter as the \textit{dominant} one, providing the state estimation for the localisation.
|
||||
Due to its robustness and good diversity the simple, more permissive filter, is then used as \textit{support} for possible sample impoverishment.
|
||||
|
||||
%kld
|
||||
This is achieved by measuring the Kullback-Leibler divergence $D_{\text{KL}}(P \|Q)$ between both filtering posterior distributions $p(\mStateVec_{t} \mid m_t, \mObsVec_{1:t})$.
|
||||
The Kullback-Leibler divergence is a non-symmetric non-negative difference between two probability distributions $Q$ and $P$.
|
||||
It is also stated as the amount of information lost when $Q$ is used to approximate $P$ \cite{Sun2013}.
|
||||
We set $D_{\text{KL}} = D_{\text{KL}}(P \|Q)$, while $P$ is the dominant and $Q$ the supporting filter.
|
||||
Since the supporting filter is more robust and ignores all environmental restrictions, we are able to make a statement whether state estimation is stuck due to sample impoverishment or not by looking at the positive exponential distribution
|
||||
As a reminder, both filters (modes) are running in parallel for the entire estimation life cycle.
|
||||
If we recognize that the dominant filter diverges from the supporting filter and thus got stuck or lost track, particles from the supporting filter will be picked with a higher probability while mixing the new particle set for the dominant filter.
|
||||
As seen before, the Markov transition matrix $\Pi_t$ provides the probability $P(m_{t+1} \mid m_t)$ for transitioning between modes.
|
||||
%In our approach those modes are the dominant graph-based filter and the supporting simple filter.
|
||||
The dominant filter's probability to draw particles from its own posterior is given by the positive exponential distribution
|
||||
%
|
||||
\begin{equation}
|
||||
f(D_{\text{KL}}, \lambda) = e^{-\lambda D_{\text{KL}}}
|
||||
\enspace .
|
||||
\label{equ:KLD}
|
||||
\end{equation}
|
||||
%
|
||||
If $D_{\text{KL}}$ increases to a certain point, \eqref{equ:KLD} provides a probability that allows for mixing the particle sets.
|
||||
Therefore, drawing particles from the support is given by $1 - f(D_{\text{KL}}, \lambda)$.
|
||||
If the Kullback-Leibler divergence $D_{\text{KL}}$ increases to a certain point, \eqref{equ:KLD} provides a probability that allows for mixing the particle sets.
|
||||
$\lambda$ depends highly on the respective filter models and is therefore chosen heuristically.
|
||||
In most cases $\lambda$ tends to be somewhere between $0.01$ and $0.10$.
|
||||
|
||||
However, it is obvious that \eqref{equ:KLD} only works reliable if the measurement noises are within reasonable limits, because the support filter depends solely on them.
|
||||
Especially Wi-Fi serves as the main source for estimation and thus attenuated or bad Wi-Fi readings are causing $D_{\text{KL}}$ to grow, even if the dominant filter provides a good position estimation.
|
||||
However, \eqref{equ:KLD} only works reliable if the measurement noise is within reasonable limits, because the support filter using the simple transition depends solely on them.
|
||||
Especially \docWIFI{} serves as the main source for estimation and thus attenuated or bad \docWIFI{} readings are causing bad estimation results for the supporting filter.
|
||||
This further leads to a growing $D_{\text{KL}}$, even if the dominant filter provides a good position estimation.
|
||||
In such scenarios a lower diversity and higher focus of the particle set, as given by the dominant filter, is required.
|
||||
We achieves this by introducing a Wi-Fi quality factor, allowing the support filter to pick particles from the dominant filter and prevent the later from doing it vice versa.
|
||||
We achieves this by introducing a \docWIFI{} quality factor, allowing the support filter to pick particles from the dominant filter and prevent the later from doing it vice versa.
|
||||
The quality factor is defined by
|
||||
%
|
||||
\begin{equation}
|
||||
@@ -150,13 +158,13 @@ The quality factor is defined by
|
||||
\label{eq:wifiQuality}
|
||||
\end{equation}
|
||||
%
|
||||
where $\bar\mRssi_\text{wifi}$ is the average of all signal-strength measurements received from the observation $\mObsVec_t$. An upper and lower bound is given by $l_\text{max}$ and $l_\text{min}$.
|
||||
|
||||
where $\bar\mRssi_\text{wifi}$ is the average of all signal strength measurements received from the observation $\mObsVec_t^{\mRssiVec_\text{wifi}}$. An upper and lower bound is given by $l_\text{max}$ and $l_\text{min}$.
|
||||
|
||||
%matrix
|
||||
To incorporate all this within the IMMPF, we utilize a non-trivial Markov switching process.
|
||||
This is done by updating the Markov transition matrix $\Pi_t$ at every time step $t$.
|
||||
As reminder, $\Pi_t$ highly influences the mixing process in \eqref{equ:immpModeMixing2}.
|
||||
Considering the above presented measures, $\Pi_t$ is two-dimensional and given by
|
||||
As reminder, $\Pi_t$ highly influences the mixing process in \eqref{equ:immpModeMixing2} and is responsible for providing the probability of moving from $m_t$ to $m_{t+1}$ in one time step.
|
||||
Considering the measures \eqref{equ:KLD} and \eqref{eq:wifiQuality} presented above, the $2$x$2$ matrix $\Pi_t$ is given by
|
||||
%
|
||||
\begin{equation}
|
||||
\Pi_t =
|
||||
@@ -169,6 +177,12 @@ Considering the above presented measures, $\Pi_t$ is two-dimensional and given b
|
||||
\end{equation}
|
||||
%
|
||||
This matrix is the centrepiece of our approach.
|
||||
It is responsible for controlling and satisfying the need of diversity and the need of focus for the whole localization approach.
|
||||
It is responsible for controlling and satisfying the need of diversity and the need of focus for the whole localisation approach.
|
||||
How $\Pi_t$ works is straightforward.
|
||||
If the dominant graph-based filter suffers from sample impoverishment (get stuck or lose track), the probability in $[\Pi_t]_{12} = (1 - f(D_{\text{KL}}))$ increases with diverging support filter.
|
||||
Consequently, the number of particles, the dominant filter draws from the supporting filter, also increases by $[\Pi_t]_{12} \cdot 100\%$.
|
||||
Similar behaviour applies to the \docWIFI{} quality factor $q(\mObsVec_t^{\mRssiVec_\text{wifi}})$.
|
||||
If the quality is low, the supporting filter regains focus by obtaining particles from the dominant's posterior.
|
||||
Therefore, recovering from sample impoverishment or degeneracy depends to a large extent on $\Pi_t$.
|
||||
|
||||
|
||||
|
||||
Reference in New Issue
Block a user