|
\section{Method name}
|
|
\label{sec:design}
|
|
|
|
\subsection{System Architecture}
|
|
We designed a thermal-box to collect the data. It has four Grideye sensors on the
|
|
corners of a 10 cm square and a Lepton 3 at the central. Figure~\ref{fig:method} shows
|
|
the system of our method. It consists four parts. The first part is to fuse multiple
|
|
Grideye image, since the resolution of singel Grideye sensor only has 64 pixels. The
|
|
second part, we train the SRCNN model with fused Grideye image
|
|
as low-resolution and downscaled Lepton 3 image as high-resolution image.
|
|
The third part, we use the
|
|
Super-resolution image to train a neural network model for recognizing current pose
|
|
is lay on back or lay on side. The last part, because of noise and the residual
|
|
heat on bed after turn over, it is difficult to figure out the current pose. We
|
|
remove the noise by median filter, and determine the current pose according to
|
|
the trend of the possibility from recognition network.
|
|
|
|
\begin{figure}[htb]
|
|
\begin{center}
|
|
\includegraphics[width=1\linewidth]{figures/method.pdf}
|
|
\caption{Illustration of Proposed Method.}
|
|
\label{fig:method}
|
|
\end{center}
|
|
\end{figure}
|
|
|
|
\subsection{Grideye Data Fusion}
|
|
|
|
On the thermal-box, there are four Grideye sensors. At the beginning, we let
|
|
the thermal-box faces to an empty bed and records the background temperature.
|
|
All the following frames will subtract this background temperature. After that,
|
|
we resize four $8 \times 8$ Grideye images to $64 \times 64$ by bilinear
|
|
interpolation and than merge them dependence on the distance between thermal-box and
|
|
bed, width of sensor square and the FOV of Grideye sensor.
|
|
|
|
\begin{enumerate}
|
|
\item $D_b$ is the distance between bed and thermal-box.
|
|
\item $D_s$ is the width of sensor square also the distance between adjacent sensors.
|
|
\item $F$ is the FOV of Grideye sensor which is about 60 degree.
|
|
\item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\frac{F}{2})})$
|
|
\end{enumerate}
|
|
|
|
\subsection{Pose determination}
|
|
|
|
We train a SRCNN model by the fused Grideye data and downscaled Lepton 3 image,
|
|
and use it to upscale all following frames to SR frames. We labeled some SR frames
|
|
to two categories. One is lay on back and the other is lay on side. Since the input
|
|
data is very small, we use a neural network consist one 2D convolution layer, one
|
|
2D max pooling, one flatten and one densely-connected layer. The possibility of
|
|
output has a very large various just after turning over because the model cannot
|
|
distinguish the residual heat on bed and the person as Figure~\ref{fig:residual_heat} shown. This
|
|
situation will slowly disappear after one or two minutes.
|
|
|
|
To determination the pose, first we use a median filter with a window size of five
|
|
to filter out the noise. Than, find the curve hull line of the upper bound and
|
|
lower bound of the data. Finally calculate the middle line of upper bound and lower bound.
|
|
Figure~\ref{fig:trend} shows the data and these lines.
|
|
|
|
We divide every data into 10 second time windows. If the middle line of the time window
|
|
is at the top one fifth, it is a lay on back. If it is at the bottom one fifth,
|
|
it is a lay on side. If the trend of line is going up, it is lay on back. Otherwise, it
|
|
is lay on side. To guarantee the confidence of result, we will only trust the pose if
|
|
there are three continuously same output.
|
|
|
|
\begin{figure}[ht]
|
|
\centering
|
|
\minipage{0.3\columnwidth}
|
|
\includegraphics[width=\linewidth]{figures/Lepton_residual_heat.png}
|
|
\caption{Residual heat on bed.}
|
|
\label{fig:residual_heat}
|
|
\endminipage
|
|
\minipage{0.65\columnwidth}
|
|
\includegraphics[width=\linewidth]{figures/MinMax.pdf}
|
|
\caption{Trend of pose.}
|
|
\label{fig:trend}
|
|
\endminipage
|
|
\end{figure}
|