Midterm Paper Example
Midterm Paper Example
RU ID:
1. Introduction
Human activity recognition is a pivotal technique in many real-world applications, such as fitness
tracking, elder care and well-being monitoring, smart home. As a traditional way to recognize activities,
wearable device-based approaches (e.g., wristband, mobile phone, body attach sensor) derive the human
motion and infer the activity through attaching motion sensors or mobile devices to human body.
However, the device attachment would bring inconvenience and discomfort to the user, especially for
elderly people with age-related memory loss. To release the users from device attachment, researchers
have been actively searching for alternative solutions, and the most promising one is device-free activity
recognition technique.
Device-free activity recognition explores wireless signals transmitted between a pair of wireless devices
(e.g., smartphone, laptop, WiFi access point) to infer the activities of a person located between or around
the sender and the receiver. The insight is that the person’s activity would introduce variations of
wireless signals reflecting activity characteristics. To date, various device-free activity recognition
approaches and systems have been developed. However, a key challenge remains unsolved, that is, the
wireless signals arriving at the receiving devices usually carry substantial information that is specific to
the environment where the activities are recorded and the human subject who conducts the activities.
Due to this reason, an activity recognition model that is trained in a specific subject in a specific
environment will not typically work well when being applied to recognize another user in another
environment.
To address this challenge, the authors of this paper proposed EI [1], a deep-learning activity recognition
framework that can remove the environment specific information entangled in the wireless signals. The
key idea is to train a domain discriminator along with a CNN based activity recognition model, which
has to capability to remove the environment specific characteristics and derive environment-
independent features. The trained CNN model can then be applied to new environment for activity
recognition. To reduce the burden of data collection, beside labeled data, the authors also propose to
use unlabeled data to train the domain discriminator. Extensive evaluations are conducted with four
different device free activity recognition testbeds, including WiFi, ultrasound, 60 GHz mmWave, and
visible light.
Based on the extracted feature, a fully collected layer is used to learn feature representations and a
SoftMax layer is used to calculate probability of over predefined activities. For the labeled training data,
the authors use cross-entropy between the prediction and the ground truth as the loss function.
Minimizing the cross-entropy during training process increases the prediction performance of the deep
learning model. In addition, for the unlabeled data, the author exploit entropy of the predictions as loss,
which can increase the model’s confidence on predicting unlabeled data.
The domain discriminator is a neural network taking the extracted feature as input and predict the
domain (i.e., environment) of input data. This seems to contradict with the goal of learning domain-
independent features of activity. But by using an adversarial loss, the feature extractor can be trained in
a way to cheat the domain discriminator (i.e., minimizing the domain prediction accuracy). Through
this minimax game, the feature extractor can extract common features in different environment. The
adversarial loss is defined as:
𝐿 = 𝐿𝛼 + 𝛼𝐿𝑢 − 𝛽𝐿𝑑 ,
where 𝐿𝛼 is the loss of labeled data and 𝐿𝑢 is the loss of unlabeled data. 𝐿𝑑 is the domain prediction
loss of the domain discriminator. By applying a negative factor to the domain loss (i.e., −𝛽), the feature
extractor will learn to cheat the domain discriminator, by minimizing the domain prediction
performance, and thus remove the domain-specific information.
3. Performance Evaluation
The authors evaluate their system on four
different device free activity recognition
testbeds, i.e., WiFi, ultrasound, 60 GHz
mmWave and visible light, to evaluate the
performance of the proposed system. To
save space we only show the performance
Figure 3: Setup for Figure 4: Setup for evaluation
of WiFi and ultrasound, with the evaluation with WiFi. with Ultrasound.
experiment setup shown in Figure 3 and
Figure 4. The authors involve 12 volunteers as subjects to conduct 6 activities, including wiping the
whiteboard, walking, moving a suitcase, rotating the chair, sitting, as well as standing up and sitting
down. The activities are conducted in 6 different rooms. They compare the proposed system with two
state-of-the-art domain adaptation methods CAT [3] and VADA [4].
WiFi. For WiFi testbed, the authors use channel state information
(CSI) measurement from 30 subcarrier groups as system input. The
CSI measurements are normalized to have zero mean and unit
standardization and denoised with Hampel filter. The processed CSI
is then segmented and fed to a deep learning model for training. The
activity recognition results under different domains (i.e.,
environment and user pair) are shown in Figure 5. The proposed deep Figure 5: Accuracy of the
learning model (EI) shows the best performance. system on CSI data.
5. Conclusion
This paper aims to address a very challenging problem, mitigating the impacts of environmental
variations in device-free activity recognition. While device-free sensing is promising, as it does not
require sensor attachment, the wireless signals are inherently susceptible to environmental dynamics
(domain variations) due to their omni-directional propagation property. Training models robust under
such dynamics is indeed an interesting and practical research problem. To achieve this goal, a domain
discriminator is used to remove environment-specific characteristics reside in wireless signals. The
underlying idea is to formulate a two-player game between the classifier (trained with labeled data from
a source domain) and the domain discriminator (trained to differentiate unlabeled data from the source
and a target domain). By applying an adversarial loss on the feature extractor, the features are optimized
to confuse the domain discriminator, making the features domain(environment)-independent.
Comprehensive evaluation on four different device-free activity recognition testbeds show the
capability of the system.
Instead of solely relying on deep learning, the system performance can be enhanced by incorporating
advanced signal processing techniques. The authors have the potential to craft state-of-the-art filtering
methods that effectively exclude signals echoed from the ambient environment, particularly from
surfaces like walls. A potential method could involve estimating the distance of a signal's propagation
path. With such an approach, the system could detect signals that reflected from a human subject and
those from walls or other static entities, and then filter them out. In addition, Sophisticated features,
grounded in advanced signal processing techniques, can be derived to provide richer information. For
example, time-frequency analysis, spectral analysis of signals, and wavelet transformation, can be
utilized to capture human motion patterns while reducing the impacts of signal reflection from static
objects. Furthermore, integrating multi-dimensional sensing measurements for activity recognition
could also potentially improve the activity recognition performance. For example, the authors may
integrate sensing measurements of multiple antennas and relationships among multiple subcarriers to
improve the activity recognition performance.
Reference
[1] Wenjun Jiang, Chenglin Miao, Fenglong Ma, Shuochao Yao, Yaqing Wang, Ye Yuan, Hongfei Xue
et al. "Towards environment independent device free human activity recognition." In Proceedings of
the 24th annual international conference on mobile computing and networking, pp. 289-304. 2018.
[2] Yaroslav Ganin and Victor Lempitsky. 2015. Unsupervised domain adaptation by backpropagation.
In International Conference on Machine Learning. 1180–1189.
[3] Mingmin Zhao, Shichao Yue, Dina Katabi, Tommi S Jaakkola, and Matt T Bianchi. 2017. Learning
sleep stages from radio signals: A conditional adversarial architecture. In International Conference on
Machine Learning. 4100–4109
[4] Rui Shu, Hung H Bui, Hirokazu Narui, and Stefano Ermon. 2018. A DIRT-T Approach to
Unsupervised Domain Adaptation. In International Conference on Learning Representations.