Browse Source

Add design and fix some package issue

git-svn-id: http://newslabx.csie.ntu.edu.tw/svn/Ginger@63 5747cdd2-2146-426f-b2b0-0570f90b98ed
master
Hobe 5 years ago
parent
commit
c1542147a5
4 changed files with 51 additions and 18 deletions
  1. +13
    -13
      trunk/RTCSA_SS/02Background.tex
  2. +21
    -1
      trunk/RTCSA_SS/03Design.tex
  3. +5
    -3
      trunk/RTCSA_SS/Main.tex
  4. +12
    -1
      trunk/RTCSA_SS/bibliography.bib

+ 13
- 13
trunk/RTCSA_SS/02Background.tex View File

@ -27,14 +27,6 @@ patches.
\end{figure} \end{figure}
\subsection{Thermal cameras}
In this work, we use two different resolution thermal camera to play the role
of low-resolution and high-resolution camera. For low-resolution camera, we use
Grid-EYE thermal camera. Grid-EYE is a thermal camera that can output
$8 \times 8$ pixels thermal data with $2.5^\circ C$ accuracy and $0.25^\circ C$
resolution at $10$ fps. For the high-resolution one, we use Lepton 3. The
specification of them are shown in Table~\ref{table:specification of devices}.
\begin{table}[hb] \begin{table}[hb]
\centering \centering
\footnotesize \footnotesize
@ -63,12 +55,20 @@ specification of them are shown in Table~\ref{table:specification of devices}.
Data Rate & 7.68 kbps & 2.34 mbps\\ Data Rate & 7.68 kbps & 2.34 mbps\\
\hline \hline
\end{tabular} \end{tabular}
\caption{Specification of Grid-EYE~\cite{grideye_datasheet} and Lepton 3~\cite{lepton_datasheet}.}
\label{table:specification of devices}
\caption{Specification of Grid-EYE and Lepton 3.}
\label{table:specificationofdevices}
\end{table} \end{table}
\subsection{Thermal cameras}
In this work, we use two different resolution thermal camera to play the role
of low-resolution and high-resolution camera. For low-resolution camera, we use
Grid-EYE thermal camera. Grid-EYE is a thermal camera that can output
$8 \times 8$ pixels thermal data with $2.5^\circ C$ accuracy and $0.25^\circ C$
resolution at $10$ fps. For the high-resolution one, we use Lepton 3. The
specification of them are shown in Table~\ref{table:specificationofdevices}.
\subsection{Related Works} \subsection{Related Works}
X. Chen et al.~\cite{7805509} proposed to use the visible camera as a guidance to super resolution the IR images. They proposed the IR-RBG multi-sensor imaging system. The approach bases on the fact that RGB channels have different correlation with IR images. To avoid wrong texture transfer, the method used cross correlation to find the proper matching between the region of IR image and RGB image. Next, the method applied sub-pixel estimation and guided filtering to remove the noise and discontinuities in IR image. At the last step of the iteration, they used truncated quadric model as the cost function to terminate the algorithm. After the iteration, the method fine tuned the output IR image with outlier detection to remove the black points which only appear at the edge of IR image.
%X. Chen et al.~\cite{7805509} proposed to use the visible camera as a guidance to super resolution the IR images. They proposed the IR-RBG multi-sensor imaging system. The approach bases on the fact that RGB channels have different correlation with IR images. To avoid wrong texture transfer, the method used cross correlation to find the proper matching between the region of IR image and RGB image. Next, the method applied sub-pixel estimation and guided filtering to remove the noise and discontinuities in IR image. At the last step of the iteration, they used truncated quadric model as the cost function to terminate the algorithm. After the iteration, the method fine tuned the output IR image with outlier detection to remove the black points which only appear at the edge of IR image.
% \begin{figure}[ht] % \begin{figure}[ht]
% \begin{center} % \begin{center}
% \includegraphics[width=0.7\linewidth]{figures/color_guided_SR.pdf} % \includegraphics[width=0.7\linewidth]{figures/color_guided_SR.pdf}
@ -77,7 +77,7 @@ X. Chen et al.~\cite{7805509} proposed to use the visible camera as a guidance t
% \end{center} % \end{center}
% \end{figure} % \end{figure}
F. Almsari et al.~\cite{8713356} applied Generative Adversarial Network (GAN) to the super-resoltuion problem called {\it TIGAN}. TIGAN consists of two networks, generator and discriminator, which learn from each other with zero-sum games and try to find an equilibrium state. The generator was trained to transfer low-resolution thermal image into super-resolution thermal image domain. The generated image should be similar to its ground truth high-resolution image domain. While the discriminator was trained to discriminate between generated image and ground truth high-resolution image. Compared to SRCNN model, it preserves the high-frequency details and gave shaper textures.
%F. Almsari et al.~\cite{8713356} applied Generative Adversarial Network (GAN) to the super-resoltuion problem called {\it TIGAN}. TIGAN consists of two networks, generator and discriminator, which learn from each other with zero-sum games and try to find an equilibrium state. The generator was trained to transfer low-resolution thermal image into super-resolution thermal image domain. The generated image should be similar to its ground truth high-resolution image domain. While the discriminator was trained to discriminate between generated image and ground truth high-resolution image. Compared to SRCNN model, it preserves the high-frequency details and gave shaper textures.
% \begin{figure}[hb] % \begin{figure}[hb]
% \begin{center} % \begin{center}
@ -88,4 +88,4 @@ F. Almsari et al.~\cite{8713356} applied Generative Adversarial Network (GAN) to
% \end{figure} % \end{figure}
% % background % % background
The assumption on the resolution of thermal images in above works is at least $120 \times 160$. Compared to the $8 \times 8$ thermal images collected by Grid-EYE sensors, more thermal features are available. In addition, these works use RGB images as reference to train the model. In this work, RGB camera will not be used so as to protect privacy.
The assumption on the resolution of thermal images in above works is at least $120 \times 160$. Compared to the $8 \times 8$ thermal images collected by Grid-EYE sensors, more thermal features are available. In addition, these works use RGB images as reference to train the model. In this work, RGB camera will not be used so as to protect privacy.

+ 21
- 1
trunk/RTCSA_SS/03Design.tex View File

@ -25,6 +25,26 @@ bed, width of sensor square and the FOV of Grideye sensor.
\item $D_b$ is the distance between bed and thermal-box. \item $D_b$ is the distance between bed and thermal-box.
\item $D_s$ is the width of sensor square also the distance between adjacent sensors. \item $D_s$ is the width of sensor square also the distance between adjacent sensors.
\item $F$ is the FOV of Grideye sensor which is about 60 degree. \item $F$ is the FOV of Grideye sensor which is about 60 degree.
\item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\feac{F}{2})})$
\item $Overlap = 64 - 64 \times (\frac{D_s}{2 \times D_b \times tan(\frac{F}{2})})$
\end{enumerate} \end{enumerate}
\subsection{Pose determination}
We train a SRCNN model by the fused Grideye data and downscaled Lepton 3 image,
and use it to upscle all following frames to SR frames. We lebeled some SR frames
to two categories. One is lay on back and the other is lay on side. Since the input
data is very small, we use a neural network consist one 2D convolution layer, one
2D max pooling, one flatten and one densely-connected layer. The possibility of
output has a very large various just after turning over because the model cannot
distinguish the residual heat on bed and the person as Figure???? showen. This
situation will slowly disappear after one or two minutes.
To determination the pose, first we use a median filter with a window size of five
to filter out the noise. Than, find the curvex hull line of the upperbound and
lowerbound of the data. Finally calculate the middle line of upperbound and lowerbound.
Figure??? and ??? shows the data and these lines.
We divide every 10 seconds data into a time window. If the middle line of the time window
is at the top one fifth, it is a lay on back. If it is at the bottom one fifth,
it is a lay on side. If the trend of line is going up, it is lay on back. Otherwise, it
is lay on side.

+ 5
- 3
trunk/RTCSA_SS/Main.tex View File

@ -13,14 +13,14 @@
%\usepackage{ntu_techrpt_cover} %\usepackage{ntu_techrpt_cover}
%\usepackage{lipsum} %\usepackage{lipsum}
%\usepackage{graphicx}
%\usepackage{times}
\usepackage{graphicx}
\usepackage{times}
%\usepackage{psfrag} %\usepackage{psfrag}
%\usepackage[tight]{subfigure} %\usepackage[tight]{subfigure}
\usepackage{setspace} \usepackage{setspace}
%\usepackage{listings} %\usepackage{listings}
%\usepackage{epsfig} %\usepackage{epsfig}
%\usepackage{longtable}
\usepackage{longtable}
%\usepackage{cases} %\usepackage{cases}
%\usepackage{subfig} %\usepackage{subfig}
\usepackage{balance} \usepackage{balance}
@ -28,6 +28,8 @@
%\usepackage{algorithm} %\usepackage{algorithm}
%\usepackage{algorithmic} %\usepackage{algorithmic}
%\usepackage[noend]{algpseudocode} %\usepackage[noend]{algpseudocode}
\usepackage{array}
\usepackage{amsmath}
%% % To add svn version number in Latex %% % To add svn version number in Latex
%\usepackage{svn-multi} %\usepackage{svn-multi}


+ 12
- 1
trunk/RTCSA_SS/bibliography.bib View File

@ -9,6 +9,7 @@ Cyber-Physical Systems},
address = {Taipei, Taiwan}, address = {Taipei, Taiwan},
publisher = {IEEE Xplore} publisher = {IEEE Xplore}
} }
@ARTICLE{ChaoDong16, @ARTICLE{ChaoDong16,
author={C. {Dong} and C. C. {Loy} and K. {He} and X. {Tang}}, author={C. {Dong} and C. C. {Loy} and K. {He} and X. {Tang}},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence}, journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
@ -20,4 +21,14 @@ Cyber-Physical Systems},
keywords={convolution;image resolution;image restoration;learning (artificial intelligence);neural nets;image super-resolution;deep learning method;end-to-end mapping;CNN;low-resolution image;color channel;deep convolutional neural network;reconstruction quality;sparse-coding;image restoration;Image resolution;Neural networks;Image reconstruction;Convolutional codes;Feature extraction;Training;Super-resolution;deep convolutional neural networks;sparse coding;Super-resolution;deep convolutional neural networks;sparse coding}, keywords={convolution;image resolution;image restoration;learning (artificial intelligence);neural nets;image super-resolution;deep learning method;end-to-end mapping;CNN;low-resolution image;color channel;deep convolutional neural network;reconstruction quality;sparse-coding;image restoration;Image resolution;Neural networks;Image reconstruction;Convolutional codes;Feature extraction;Training;Super-resolution;deep convolutional neural networks;sparse coding;Super-resolution;deep convolutional neural networks;sparse coding},
doi={10.1109/TPAMI.2015.2439281}, doi={10.1109/TPAMI.2015.2439281},
ISSN={0162-8828}, ISSN={0162-8828},
month={Feb},}
month={Feb},}
@misc{lepton_datasheet,
title = {FLIR LEPTON® 3 Long Wave Infrared (LWIR) Datasheet},
howpublished = {\url{https://media.digikey.com/pdf/Data\%20Sheets/FLIR\%20PDFs/Lepton3_LWIR_DS_Rev100_4-4-17.pdf}},
}
@misc{grideye_datasheet,
title = {Infrared Array Sensor Grid-EYE (AMG88) Datasheet},
howpublished = {https://datasheet.octopart.com/AMG8833-Panasonic-datasheet-62338626.pdf},
}

Loading…
Cancel
Save