555
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 

78 lines
3.7 KiB

\section{System Architecture}
\label{sec:design}
We designed a thermal-box to collect the data. It has four Grideye sensors on the
corners of a 10 cm square and a Lepton 3 at the central. Figure~\ref{fig:method} shows
the system of our method. It consists four parts. The first part is to fuse multiple
data from Grideye sensors into a low-resolution image, since the resolution of a
single Grideye sensor too low to make a decision. The
second part, we train the SRCNN model with fused Grideye image
as low-resolution and downscaled Lepton 3 image as high-resolution image.
The third part, we use the
Super-resolution image to train a neural network model for recognizing current pose
is lay on back or lay on side. The last part, to reduce the noise and effect cause by
the residual heat on bed after turning over. We
remove the noise by median filter, and determine the current pose according to
the trend of the possibility from recognition network.
\begin{figure}[htb]
\begin{center}
\includegraphics[width=1\linewidth]{figures/method.pdf}
\caption{Illustration of Proposed Method.}
\label{fig:method}
\end{center}
\end{figure}
\subsection{Grideye Data Fusion}
On the thermal-box, there are four Grideye sensors. At the beginning, we let
the thermal-box faces to an empty bed and records the background temperature.
All the following frames will subtract this background temperature. After that,
we resize four $8 \times 8$ Grideye images to $64 \times 64$ by bilinear
interpolation and then merge them dependence on the distance between thermal-box and
bed, distance between sensors and the FOV of Grideye sensor. In our case, $D_B$ is
150 cm, and $D_s$ is 10 cm.
\begin{enumerate}
\item $D_b$ is the distance between bed and thermal-box.
\item $D_s$ is the width of sensor square also the distance between adjacent sensors.
\item $F$ is the FOV of Grideye sensor which is about 60 degree.
\item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\frac{F}{2})})$
\end{enumerate}
\subsection{Turning Over Determination}
We train a SRCNN model by the fused Grideye image and downscaled Lepton 3 image,
and use it to enhance all following Grideye frames to SR frames. We labeled some SR frames
into two categories, lay on back and lay on side. Since the input
data is very small, we use a neural network consist one 2D convolution layer, one
2D max pooling, one flatten and one densely-connected layer. The possibility of
output has a very large various just after turn over because the model cannot
distinguish the residual heat on bed and the person as Figure~\ref{fig:residual_heat} shown. This
situation will slowly disappear after one or two minutes.
To determination the pose, first we use a median filter with a window size of five
to filter out the noise. Then, find the curve hull line of the upper bound and
lower bound of the data. Finally, calculate the middle line of upper bound and
lower bound, and regrad it as the trend of the pose changing. Figure~\ref{fig:trend}
shows the filitered data and these lines.
We divide every data into 10 second time windows. If the middle line of the time window
is at the top one fifth, or the trend is going up, it is a lay on back. If it is at the
bottom one fifth, or the trend is going down, it is a lay on side. If there are three
continuously same poses, and different from the last turning over, it will be count as
another turning over.
\begin{figure}[ht]
\centering
\minipage{0.3\columnwidth}
\includegraphics[width=\linewidth]{figures/Lepton_residual_heat.png}
\caption{Residual heat on bed.}
\label{fig:residual_heat}
\endminipage
\minipage{0.65\columnwidth}
\includegraphics[width=\linewidth]{figures/MinMax.pdf}
\caption{Trend of pose.}
\label{fig:trend}
\endminipage
\end{figure}