- \section{Method name}
 - \label{sec:design}
 - 
 - \subsection{System Architecture}
 - We designed a thermal-box to collect the data. It has four Grideye sensors on the
 - corners of a 10 cm square and a Lepton 3 at the central. Figure~\ref{fig:method} shows
 - the system of our method. It consists four parts. The first part is to fuse multiple
 - Grideye image, since the resolution of singel Grideye sensor only has 64 pixels. The
 - second part, we train the SRCNN model with fused Grideye image
 - as low-resolution and downscaled Lepton 3 image as high-resolution image.
 - The third part, we use the
 - Super-resolution image to train a neural network model for recognizing current pose
 - is lay on back or lay on side. The last part, because of noise and the residual
 - heat on bed after turn over, it is difficult to figure out the current pose. We
 - remove the noise by median filter, and determine the current pose according to
 - the trend of the possibility from recognition network.
 - 
 - \begin{figure}[htb]
 - 	\begin{center}
 - 		\includegraphics[width=1\linewidth]{figures/method.pdf}
 - 		\caption{Illustration of Proposed Method.}
 - 		\label{fig:method}
 - 	\end{center}
 - \end{figure}
 - 
 - \subsection{Grideye Data Fusion}
 - 
 - On the thermal-box, there are four Grideye sensors. At the beginning, we let
 - the thermal-box faces to an empty bed and records the background temperature.
 - All the following frames will subtract this background temperature. After that,
 - we resize four $8 \times 8$ Grideye images to $64 \times 64$ by bilinear
 - interpolation and than merge them dependence on the distance between thermal-box and
 - bed, width of sensor square and the FOV of Grideye sensor.
 - 
 - \begin{enumerate}
 - \item $D_b$ is the distance between bed and thermal-box.
 - \item $D_s$ is the width of sensor square also the distance between adjacent sensors.
 - \item $F$ is the FOV of Grideye sensor which is about 60 degree.
 - \item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\frac{F}{2})})$
 - \end{enumerate}
 - 
 - \subsection{Pose determination}
 - 
 - We train a SRCNN model by the fused Grideye data and downscaled Lepton 3 image,
 - and use it to upscale all following frames to SR frames. We labeled some SR frames
 - to two categories. One is lay on back and the other is lay on side. Since the input
 - data is very small, we use a neural network consist one 2D convolution layer, one
 - 2D max pooling, one flatten and one densely-connected layer. The possibility of
 - output has a very large various just after turning over because the model cannot
 - distinguish the residual heat on bed and the person as Figure~\ref{fig:residual_heat} shown. This
 - situation will slowly disappear after one or two minutes.
 - 
 - To determination the pose, first we use a median filter with a window size of five
 - to filter out the noise. Than, find the curve hull line of the upper bound and
 - lower bound of the data. Finally calculate the middle line of upper bound and lower bound.
 - Figure~\ref{fig:trend} shows the data and these lines.
 - 
 - We divide every data into 10 second time windows. If the middle line of the time window
 - is at the top one fifth, it is a lay on back. If it is at the bottom one fifth,
 - it is a lay on side. If the trend of line is going up, it is lay on back. Otherwise, it
 - is lay on side. To guarantee the confidence of result, we will only trust the pose if
 - there are three continuously same output.
 - 
 - \begin{figure}[ht]
 -   \centering
 -   \minipage{0.3\columnwidth}
 -     \includegraphics[width=\linewidth]{figures/Lepton_residual_heat.png}
 -     \caption{Residual heat on bed.}
 -     \label{fig:residual_heat}
 -   \endminipage
 -   \minipage{0.65\columnwidth}
 -     \includegraphics[width=\linewidth]{figures/MinMax.pdf}
 -     \caption{Trend of pose.}
 -     \label{fig:trend}
 -   \endminipage
 - \end{figure}
 
 
  |