\section{WiFi Location Estimation} \label{sec:optimization} The \docWIFI{} sensor infers the pedestrian's current location based on a comparison between recent measurements (the smartphone continuously scans for nearby \docAP{}s) and reference measurements or signal strength predictions for well known locations: \begin{equation} p(\vec{o}_t \mid \vec{q}_t)_\text{wifi} = p(\mRssiVecWiFi \mid \mPosVec) = \prod_{\mRssi_{i} \in \mRssiVec{}} p(\mRssi_{i} \mid \mPosVec),\enskip %\mPos = (x,y,z)^T \mPosVec \in \R^3 \label{eq:wifiObs} \end{equation} % \begin{equation} p(\mRssi_i \mid \mPosVec) = \mathcal{N}(\mRssi_i \mid \mu_{i,\mPosVec}, \sigma_{i,\mPosVec}^2) \label{eq:wifiProb} \end{equation} In \refeq{eq:wifiProb}, $\mu_{i,\mPosVec}$ and $\sigma_{i,\mPosVec}$ denote the average signal strength and corresponding standard deviation for the \docAPshort{} identified by $i$, that should be measurable given the location $\mPosVec = (x,y,z)^T$. Those two value can be determined using various methods. Most common, as of today, seems fingerprinting, where hundreds of locations throughout the building are scanned beforehand. The received \docAP{}s including their average signal strength and deviation denote each location's fingerprint \cite{radar}. % While allowing for highly accurate location estimations, given enough fingerprints, such a setup is costly, as fingerprinting is a manual process. % We therefore use a model to predict the average signal strength for each location, based on the \docAPshort{}'s position $\mPosAPVec{} = (x,y,z)^T$ and a few additional parameters. \subsection{Signal Strength Prediction Model} \begin{equation} \mRssi = \mTXP{} + 10 \mPLE{} + \log_{10} \frac{d}{d_0} + \mGaussNoise{} \label{eq:logDistModel} \end{equation} The log distance model \cite{TODO} in \refeq{eq:logDistModel} is a commonly used signal strength prediction model that is intended for line-of-sight predictions. However, depending on the surroundings, the model is versatile enough to also serve for indoor purposes. % It predicts an \docAP{}'s signal strength for an arbitrary location $\mPosVec{}$ given the distance between both and two environmental parameters: The \docAPshort{}'s signal strength \mTXP{} measurable at a known distance $d_0$ (usually \SI{1}{\meter}) and the signal's depletion over distance \mPLE{}, which depends on the \docAPshort{}'s surroundings like walls and other obstacles. \mGaussNoise{} is a zero-mean Gaussian noise and models the uncertainty. As \mPLE{} depends on the architecture around the transmitter, the model is bound to homogenous surroundings like one floor, solely divided by drywalls of the same thickness and material. % The log normal shadowing model is a slight modification, to adapt the log distance model to indoor use cases. It introduces an additional parameter, that considers obstacles between (line-of-sight) the \docAPshort{} and the location in question by attenuating the signal with a constant value. % Depending on the use case, this value describes the number and type of walls, ceilings, floors etc. between both positions. For obstacles, this requires an intersection-test of each obstacle with the line-of-sight, which is costly for larger buildings. For real-time use on a smartphone, a (discretized) model pre-computation might thus be necessary \todo{cite competition}. Furthermore this requires a detailed floorplan, that includes material information for walls, doors, floors and ceilings. Throughout this work, we thus use a tradeoff between both models, where walls are ignored and only floors/ceilings are considered. Assuming buildings with even floor levels, the number of floors/ceilings between two position can be determined without costly intersection checks and thus allows for real-time use cases. \begin{equation} x = \mTXP{} + 10 \mPLE{} + \log_{10} \frac{d}{d_0} + \numFloors{} \mWAF{} + \mGaussNoise{} \label{eq:logNormShadowModel} \end{equation} In \refeq{eq:logNormShadowModel}, those are included using a constant attenuation factor \mWAF{} multiplied by the number of floors/ceilings \numFloors{} between sender and the location in question. The attenuation \mWAF{} (per element) depends on the building's architecture and for common, steel enforced concrete floors $\approx 8.0$ is a viable choice \cite{TODO}. \subsection {Model Parameters} As previously mentioned, for the prediction model to work, one needs to know the location $\mPosAPVec_i$ for every permanently installed \docAP{} $i$ within the building to derive the distance $d$, plus its environmental parameters \mTXP{}, \mPLE{} and \mWAF{}. While it is possible to use empiric values for those environmental parameters \cite{Ebner-15}, the positions are mandatory. For many buildings, there should be floorplans that include the locations of all installed transmitters. If so, a model setup takes only several minutes to (vaguely) position the \docAPshort{}s within a virtual map and assigning them some fixed, empirically chosen parameters for \mTXP{}, \mPLE{} and \mWAF{}. Depending on the building's architecture this might already provide enough accuracy for some use-cases, where a vague location information is sufficient. \subsection{Model Parameter Optimization} For systems that demand a higher accuracy, one can choose a compromise between fingerprinting and pure empiric model parameters where (some) model parameters are optimized, based on a few reference measurements throughout the building. Obviously, the more parameters are unknown ($\mPosAPVec{}, \mTXP{}, \mPLE{}, \mWAF{}$) the more reference measurements are necessary to provide a stable optimization. Depending on the desired accuracy, setup time and whether the transmitter positions are known or unknown, several optimization strategies arise, where not all 6 parameters are optimized, but only some of them. Just optimizing \mTXP{} and \mPLE{} with constant \mWAF{} and known transmitter position usually means optimizing a convex function as can be seen in figure \ref{fig:wifiOptFuncTXPEXP}. For such error functions, algorithms like gradient descent \cite{TODO} and (downhill) simpelx \cite{TODO} are well suited and will provide the global minima: \begin{equation} argmin_{bla} blub() TODO TODO TODO \end{equation} \begin{figure} \input{gfx/wifiop_show_optfunc_params} \label{fig:wifiOptFuncTXPEXP} \caption{ The average error (in \SI{}{\decibel}) between all reference measurements and corresponding model predictions for one \docAPshort{} dependent on \docTXP{} \mTXP{} and \docEXP{} \mPLE{} [known position $\mPosAPVec{}$, fixed \mWAF{}] denotes a convex function. } \end{figure} However, optimizing an unknown transmitter position usually means optimizing a non-convex, discontinuous function, especially when the $z$-coordinate, that influences the number of attenuating floors/ceilings, is involved. While the latter can be mitigated by introducing a continuous function for the number $n$ of floors/ceilings, like a sigmoid, the function is not necessarily convex. As can be seen in figure \ref{fig:wifiOptFuncPosYZ}, there are two local minima and only one of both also is a global one. \begin{figure} \input{gfx/wifiop_show_optfunc_pos_yz} \label{fig:wifiOptFuncPosYZ} \caption{ The average error (in \SI{}{\decibel}) between reference measurements and model predictions for one \docAPshort{} dependent on $y$- and $z$-position [fixed $x$, \mTXP{}, \mPLE{} and \mWAF{}] usually denotes a non-convex function with multiple [here: two] local minima. } \end{figure} Such functions demand for optimization algorithms, that are able to deal with non-convex functions, like genetic approaches. However, initial tests indicated that while being superior to simplex and similar algorithms, the results were not satisfactorily and the optimization often did not converge. As the Range of the six to-be-optimized parameters is known ($\mPosAPVec{}$ within the building, \mTXP{}, \mPLE{}, \mWAF{} within a sane interval), we used some modifications. The algorithms initial population is uniformly sampled from the known range. During each iteration the best \SI{25}{\percent} of the population are kept and the remaining entries are re-created by modifying the best entries with uniform random values within $\pm$\SI{10}{\percent} of the known range. To stabilize the result, the allowed modification range (starting at \SI{10}{\percent}) is reduced over time, known as cooling \cite{todo}. \subsection{Modified Signal Strength Model} %\todo{nicht: during initial eval, sondern gleich sagen, dass die vermutung nahe liegt, dass das modell %nicht gut klappen wird, weil waende und unser metall-glas nicht beruecksichtigt werden. deshalb %versuchen wir ein anderes modell das immernoch live arbeiten kann} %During the initial eval, some issues were discovered. While aforementioned optimization was able to %reduce the error between reference measurements and model estimations to \SI{50}{\percent}, %the position estimation \ref{eq:wifiProb} did not benefit from improved model parameters. %To the contrary, there were several situations throughout the testing walks, where %the inferred location was more erroneous than before. As the used model tradeoff does not consider walls, it is expected to provide erroneous values for regions that are heavily shrouded by e.g. steel-enforced concrete or metallised glass. \subsection{\docWIFI{} quality factor} Past evaluations showed, that there are many situations where the \docWIFI{} location estimation is highly erroneous. Either when the signal strength prediction model does not match real world conditions or the received measurements are ambiguous and there is more than one location within the building that matches those readings. Both cases can occur e.g. in areas surrounded by concrete walls where the model does not match the real world conditions as those walls are not considered, and the smartphone barely receives some \docAPshort{}s due to the high attenuation. If such a sensor error occurs only for a short time period, the recursive density estimation \refeq{eq:recursiveDensity} is able to compensate those errors using other sensors and the movement model. However, if the error persists for a longer time period, the error will slowly distort the posterior distribution. As our movement model depends on the actual floorplan, the density might get trapped e.g. within a room if the other sensors are not able to compensate for the \docWIFI{} error. Thus, we try to determine the quality of received \docWIFI{} measurements, which allows for temporarily disabling \docWIFI{}'s contribution within the evaluation \refeq{eq:evalDensity} for situations where the quality is insufficient. In \refeq{eq:wifiQuality} we use the average signal strength of all \docAP{}s seen within one measurement and scale this value to match a region of $[0, 1]$ depending on an upper- and lower bound. If the returned quality falls below a certain threshold, \docWIFI{} is ignored within the evaluation. \begin{equation} \newcommand{\leMin}{l_\text{min}} \newcommand{\leMax}{l_\text{max}} q(\mRssiVec) = \max(0, \min( \frac{ \bar\mRssi - \leMin }{ \leMax - \leMin }, 1 ) ) \label{eq:wifiQuality} \end{equation} \subsection {VAP grouping} \label{sec:vap} Assuming normal conditions, the received signal strength at one location will also (strongly) vary due to environmental conditions like temperature, humidity, open/closed doors and RF interference. Fast variations can be addressed by averaging several consecutive measurements at the expense of a delay in time. To prevent this delay we use the fact, that many buildings use so called virtual access points where one physical hardware \docAP{} provides more than one virtual network to connect to. They can usually be identified, as only the last digit of the MAC-address is altered among the virtual networks. % As those virtual networks normally share the same frequency, they are unable to transmit at the same time. When scanning for \docAPshort{}s one will thus receive several responses from the same hardware, all with a very small delay in time (micro- to milliseconds). Such measurements may be grouped using some aggregate function like average, median or maximum. wie wird optimiert a) bekannte pos + empirische params b) bekannte pos + opt params (fur alle APs gleich) [simplex] c) bekannte pos + opt params (eigene je AP) [simplex] d) alles opt: pos und params (je ap) [range-random] optimierung ist tricky. auch wegen dem WAF der ja sprunghaft dazu kommt, sobald messung und AP in zwei unterschiedlichen stockwerken liegen.. und das selbst wenn hier vlt sichtkontakt möglich wäre, da der test 2D ist und nicht 3D aps sind (statistisch) unaebhaengig. d.h., jeder AP kann fuer sich optimiert werden. optimierung des gesamtsystems ist nicht notwendig. pro AP also 6 params. pos x/y/z, txp, exp, waf