You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 

76 lines
3.6 KiB

\section{Method name}
\label{sec:design}
\subsection{System Architecture}
We designed a thermal-box to collect the data. It has four Grideye sensors on the
corners of a 10 cm square and a Lepton 3 at the central. Figure~\ref{fig:method} shows
the system of our method. It consists four parts. The first part is to fuse multiple
Grideye image, since the resolution of singel Grideye sensor only has 64 pixels. The
second part, we train the SRCNN model with fused Grideye image
as low-resolution and downscaled Lepton 3 image as high-resolution image.
The third part, we use the
Super-resolution image to train a neural network model for recognizing current pose
is lay on back or lay on side. The last part, because of noise and the residual
heat on bed after turn over, it is difficult to figure out the current pose. We
remove the noise by median filter, and determine the current pose according to
the trend of the possibility from recognition network.
\begin{figure}[htb]
\begin{center}
\includegraphics[width=1\linewidth]{figures/method.pdf}
\caption{Illustration of Proposed Method.}
\label{fig:method}
\end{center}
\end{figure}
\subsection{Grideye Data Fusion}
On the thermal-box, there are four Grideye sensors. At the beginning, we let
the thermal-box faces to an empty bed and records the background temperature.
All the following frames will subtract this background temperature. After that,
we resize four $8 \times 8$ Grideye images to $64 \times 64$ by bilinear
interpolation and than merge them dependence on the distance between thermal-box and
bed, width of sensor square and the FOV of Grideye sensor.
\begin{enumerate}
\item $D_b$ is the distance between bed and thermal-box.
\item $D_s$ is the width of sensor square also the distance between adjacent sensors.
\item $F$ is the FOV of Grideye sensor which is about 60 degree.
\item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\frac{F}{2})})$
\end{enumerate}
\subsection{Pose determination}
We train a SRCNN model by the fused Grideye data and downscaled Lepton 3 image,
and use it to upscale all following frames to SR frames. We labeled some SR frames
to two categories. One is lay on back and the other is lay on side. Since the input
data is very small, we use a neural network consist one 2D convolution layer, one
2D max pooling, one flatten and one densely-connected layer. The possibility of
output has a very large various just after turning over because the model cannot
distinguish the residual heat on bed and the person as Figure~\ref{fig:residual_heat} shown. This
situation will slowly disappear after one or two minutes.
To determination the pose, first we use a median filter with a window size of five
to filter out the noise. Than, find the curve hull line of the upper bound and
lower bound of the data. Finally calculate the middle line of upper bound and lower bound.
Figure~\ref{fig:trend} shows the data and these lines.
We divide every data into 10 second time windows. If the middle line of the time window
is at the top one fifth, it is a lay on back. If it is at the bottom one fifth,
it is a lay on side. If the trend of line is going up, it is lay on back. Otherwise, it
is lay on side. To guarantee the confidence of result, we will only trust the pose if
there are three continuously same output.
\begin{figure}[ht]
\centering
\minipage{0.3\columnwidth}
\includegraphics[width=\linewidth]{figures/Lepton_residual_heat.png}
\caption{Residual heat on bed.}
\label{fig:residual_heat}
\endminipage
\minipage{0.65\columnwidth}
\includegraphics[width=\linewidth]{figures/MinMax.pdf}
\caption{Trend of pose.}
\label{fig:trend}
\endminipage
\end{figure}