Angle Insensitive Human Motion and Posture Recognition Based on 4D Imaging Radar and Deep Learning Classifiers
Angle Insensitive Human Motion and Posture Recognition Based on 4D Imaging Radar and Deep Learning Classifiers
DOI
10.1109/JSEN.2022.3175618
Publication date
2022
Document Version
Accepted author manuscript
Published in
IEEE Sensors Journal
Citation (APA)
Zhao, Y., Yarovoy, A., & Fioranelli, F. (2022). Angle-insensitive Human Motion and Posture Recognition
Based on 4D imaging Radar and Deep Learning Classifiers. IEEE Sensors Journal, 22(12), 12173-12182.
https://doi.org/10.1109/JSEN.2022.3175618
Important note
To cite this publication, please use the final published version (if applicable).
Please check the document version above.
Copyright
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent
of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Takedown policy
Please contact us and provide details if you believe this document breaches copyrights.
We will remove access to the work immediately and investigate your claim.
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
2 IEEE SENSORS JOURNAL, VOL. XX, NO. XX, XXXX 2017
Doppler information suffer from two inherent limitations. proposed pipeline also shows robust results in case of low
Firstly, the classified movements are typically artificially lim- signal-to-noise ratio (SNR), varying dimensions of the virtual
ited to be performed along the line-of-sight direction of the apertures (i.e., the number of array channels that worsen or
radar, or at a small aspect angle so that the Doppler infor- improve angular resolutions), and leave-one-subject-out test
mation remains representative enough. The second limitation to validate performance for unseen individuals.
is that static postures on their own or in between continuous The rest of the paper is organized as follows. Section II de-
sequences of movements are rarely investigated, since their scribes the proposed method. Section III presents the measured
micro-Doppler signatures are not easily distinguishable from dataset for validating the performance of the proposed method.
the static clutter. Section IV discusses the attained results for the proposed
To overcome the first limitation, the usage of distributed method and its comparison with the state of the art. Finally,
or even multistatic radar systems has been proposed, in order conclusions are drawn in Section V.
to simultaneously sample and reconstruct the micro-Doppler
signatures from different aspect angles [18]–[22]. While effec- II. D ESCRIPTION OF THE P ROPOSED P IPELINE AND
tive, this approach requires the usage of multiple nodes, with C OMPARATIVE B ASELINES
an increase in complexity of the overall system and the need to
cope with the synchronization of data from the different sensor Radar conventionally used in HAR (i.e., radar with a single
nodes. Another approach, recently proposed thanks to the receiver and operating at the relatively low carrier frequency
availability of millimeter-wave (mm-wave) MIMO (Multiple in the 5.8 or 24 GHz ISM bands) generates data from which
Input Multiple Output) radars, is based on the use of 4D information related to four intrinsic features of the object
imaging radar that exploits azimuth and elevation resolution can be extracted: range as 1D spatial information, Doppler
capabilities to attain additional spatial information on the proportional to the target’s radial velocity, received power
subject’s posture [23], [24]. While promising, this approach proportional to the Radar Cross Section of the object, and
is still not investigated in detail in the literature and there is a the temporal relations from the movements of the body parts.
scope to define effective classification processing pipelines that The usage of these four features or representations of the data
can take the advantages of mm-wave 4D imaging capabilities has been thoroughly analyzed in the literature, and it appears
for HAR. that the state-of-the-art research mostly relies on the Doppler
To address the aforementioned radar-based HAR issues of information [25]–[28].
unfavorable orientations (in terms of aspect angles with respect To overcome the inherent limitations of Doppler informa-
to radar’s line-of-sight) and static human postures, this work tion, additional intrinsic features must be introduced. 4D imag-
proposes a radar-based classification pipeline that exploits the ing radar at mm-wave frequencies can provide an estimation
richer information provided by mm-wave 4D imaging radar. of the spatial occupancy of the human body in height and
The main contributions of the proposed pipeline are as follows. width in different positions. Specifically, the usage of multiple
channels in an antenna array allows estimating the angles of
• Unlike in the other studies such as [23] where the arrival of the targets. At mm-wave frequencies, human bodies
radar point clouds are treated as images, the proposed are perceived as extended targets, with multiple scatterers
pipeline is designed with the goal to exploit all the six generated by each moving body part forming the so-called
intrinsic ’features’ obtained by imaging radar, namely point clouds (PCs) [23]. This scattering behavior, combined
range, azimuth, elevation, Doppler, received power, and with the angular estimation capabilities on both azimuth and
time information, rather than focusing on just one specific elevation, enables a new, broader ’feature space’ to explore
data representation. for radar-based HAR. However, when operating at mm-wave,
• The hierarchical structure of the proposed pipeline sim- a disadvantage to account for is that the detection range is
plifies the task of HAR in a multi-angle scenario, with shorter than at lower frequencies conventionally used for HAR
the help of the designated neural networks, T-Net, and such as the ISM bands of 2.4, 5.8, and 24 GHz, because of
achieves classification of both static postures and dynamic the higher propagation losses. Nevertheless, current mm-wave
motions. systems at 60-77 GHz show good detection and classification
• The pipeline is designed to be robust to a noisy and capabilities at the ranges of interest for HAR applications.
limited amount of radar data by replacing the Max Pool- The overview of the proposed method to exploit all these
ing layer in T-Net and PointNet with Average Pooling, features is given in Figure 1. Specifically, this method exploits
and deliberately using lighter-weight neural networks, the six intrinsic features of range, azimuth, elevation, Doppler,
respectively. received power, and time by combining both PCs and spec-
To validate the performance of the proposed pipeline, a custom trograms as the input data representations. The hierarchical
experimental dataset is collected including 8 human subjects structure of the classification pipeline includes the so-called
performing 6 in-place activities (including 4 motions and 2 ’orientation classification module’ to first classify to which
postures) at 5 different orientations. The measurement was orientation the human subject is facing toward (e.g., 0, 45,
conducted in an office-like room to simulate an indoor real-life 90, 135, 180◦ ). Then, based on the predictions made by this
environment. The proposed pipeline attains the classification module, the ’PC classification module’ predicts which posture
accuracy of 87.1%, which is significantly higher than the or motion pair the input belongs to (more details about the
state-of-the-art alternatives applied to the same dataset. The definition of motion pair is given later in this section). Finally,
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
AUTHOR et al.: IMAGING RADAR-BASED HUMAN ACTIVITY RECOGNITION 3
Fig. 1. Overview of the proposed classification pipeline, where the main contributions are the parallel processing and fusion of PCs and
spectrograms, and the usage of T-Net to obtain angular orientation insensitivity. Besides the data generation and pre-processing operations (top
box), the three main modules include the orientation classification, the PC classification, and the spectrogram classification.
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
4 IEEE SENSORS JOURNAL, VOL. XX, NO. XX, XXXX 2017
classification problem is simplified to be a uni-angle the proposed pipeline and is named as baseline-1.
problem. Furthermore, additional comparisons are made with other
3) For each predicted orientation of the human subject, the pioneering studies investigating the feasibility of applying
PC classification module is then used. This module imaging radar for radar-based HAR. In particular, [23] used
is adapted from PointNet [29] by proposing a new ’snapshots’ of PCs aggregated over the interval of activity as
global symmetric function average pooling to replace the input to their designed deep convolutional neural network.
the original max pooling in order to be robust to Their method essentially treats imaging radar-based HAR as a
noise. The PC classification module makes a prediction 2D image classification problem, unlike the proposed pipeline
according to the spatial distribution of PCs aggregated that processes the 3D coordinates of the detected PCs. It is
over 20 frames. As previously mentioned, without this crucial to determine whether the full use of all the intrinsic
aggregation, the sparsity of the radar PC over one or few features provided by 4D imaging radar helps to achieve better
frames would make the classification task for human HAR performance, so [23] is used as a replica of Orientation
movements or postures too challenging. An undesired classification module, PC classification module and the entire
consequence of this aggregation over time is that dy- pipeline for comparison. This comparative architecture is
namic activities and motions that are implicitly shorter named baseline-2. Furthermore, since the classifier in [23] is
than the duration of 20 frames would result in very sim- not explicitly compared with other image-based classifiers in
ilar PCs, for example sitting down & standing up from the literature, four of them - namely, VGG [32], ResNet [33],
sitting, or bending over & standing up from bending. DenseNet [34] and ViT [35] - are also used for comparison in
Therefore, the output of this module may include classes this paper. They are named baseline-3 to baseline-6.
that are possible motion pairs (e.g., (a) sitting down on a
chair or standing up from sitting, and (b) bending over or III. DATASET D ESCRIPTION
standing up from bending), or static postures such as (c) To validate the performance of the proposed pipeline, a mm-
sitting still and (d) standing still. These are summarized wave MIMO FMCW radar developed by Texas Instruments
in Table II. (cascaded AWR2243 radar) was used to collect an experimen-
4) The last module is the spectrogram classification mod- tal dataset. Mm-wave FMCW MIMO radar has been a popular
ule. This module takes the predicted class from the choice in short-range applications such as indoor HAR thanks
previous module and uses the spectrogram as the input to its flexibility, low cost, and small physical size as an off-the-
to the AlexNet [30] to recognize the individual motion shelf product, mostly driven by the technological development
class within the motion pair (a) or (b). The choice of in the automotive sector. Specifically, the cascaded AWR2243
AlexNet [30] is due to its relatively simple structure and radar operates at 79GHz with 12 transmitters and 16 receivers.
easier convergence compared to deeper neural networks, Using MIMO configuration, the radar produced a virtual array
but it theoretically can be replaced by other spectrogram- with an aperture of 43λ × 3λ (see Figure 2), where λ is the
based classification approaches from the literature [25], center wavelength of the transmitted signal. For the purpose
[26] if desired. Therefore, the final classification output of comparative studies, we mostly inherited the FMCW wave-
of the proposed pipeline is a combination of modules form parameters from [23], which are given as follows: chirp
using both PC and spectrograms as input data represen- duration 63 µsec, chirp slope 60 MHz/µsec, chirps per frame
tations. 128, frame period 100 ms, frequency bandwidth 2.84 GHz, and
A/D sampling rate 2.7 MHz. The radar specifications derived
As claimed by [31], an angle-insensitive HAR pipeline from these parameters are as follows: range resolution of 52.8
should be robust to achieve angle-insensitive HAR given mm, azimuth resolution of 1.4 degrees (at broadside), elevation
training data collected at multiple orientations, or even ideally resolution of 18 degrees (at broadside), velocity resolution
given training data collected at one orientation only. Thus, two of ±0.0286 m/s. These resolutions are expected to provide
definitions are given for how to use a classification pipeline in sufficient information on human dynamics as well as body
terms of different combinations of training data with respect shapes.
to human orientation. These include training with data from
one human orientation and testing with multiple orientations, TABLE II
which is termed as SAC (Single Angle Classifier), and MAC L IST OF MOTIONS , MOTION PAIRS , AND POSTURES FOR THE MEASURED
(Multiple Angle Classifier) if training is performed with data DATASET. M OTION PAIRS AND POSTURES REPRESENT THE
INTERMEDIATE OUTPUT CLASSES ( A - B - C - D ) OF THE PC
collected at multiple orientations. It should be noted that
CLASSIFICATION MODULE , WHEREAS INDIVIDUAL MOTIONS AND
the orientation classification module becomes insignificant for POSTURES REPRESENT THE FINAL OUTPUT CLASSES OF THE
SAC cases, and thus is bypassed in the proposed pipeline. PROPOSED PIPELINE (1-6).
Angle Sensitivity Matrix (ASM) and Angle Sensitivity Vec-
Motion
tor (ASV) are used as metrics to evaluate the classification Motion Posture
pair
performance of SAC and MAC, respectively, as proposed in 1. Sitting down c 5. Sitting still
[31]. The former has two dimensions of training and test a
2. Standing up from sitting d 6. Standing still
orientations, and the latter is compacted from a matrix into a 3. Bending over
b
vector, as data of all orientations are used for training together. 4. Standing up from bending
The work in [31] is used as a comparative baseline to evaluate
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
AUTHOR et al.: IMAGING RADAR-BASED HUMAN ACTIVITY RECOGNITION 5
noise to raw data, the final obtained SNR values are 20dB,
18dB, 15dB, 13dB, 10dB and 8dB. Moreover, ”small-aperture
datasets” were generated by selecting the raw data of only
a subset of virtual channels for subsequent signal processing
(essentially the subarray-1 to 4 shown in Figure 2). This was
done to test the effect of reduced angular resolutions on the
generated PCs and on the subsequent classification.
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
6 IEEE SENSORS JOURNAL, VOL. XX, NO. XX, XXXX 2017
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
AUTHOR et al.: IMAGING RADAR-BASED HUMAN ACTIVITY RECOGNITION 7
TABLE VI
Q UANTITATIVE METRICS INSPIRED FROM THE baseline-1 [31].
C OMPARISON OF RESULTS USING THE PROPOSED PIPELINE AND THEIR
CLASSIFICATION APPROACH .
scores for the different baselines from the literature [23], [32]–
[35] and the proposed pipeline. The proposed pipeline shows
a lead 12.2% in accuracy and 13.1% in F1-score compared
with baseline-2 to 6 when classifying 6 activities end-to-end.
Furthermore, the accuracy advantage of the proposed pipeline
is at least 22.4% for classifying human orientations and at least
11.5% when the image classifier baselines are used as a replica
of the proposed PC classification module. All these results Fig. 5. Classification accuracy of the proposed pipeline and two
selected baselines with respect to varying number of training samples.
appear to suggest that fully exploiting all the information
provided by an imaging radar, e.g. using together spectrograms
and the PC coordinates as input to classifiers, is more helpful shrinks sharply from 87.1% to 74.0% accuracy for 6 activities
in achieving good classification results than treating PCs as and from 83.0% to 73.5% for 4 motions, as the number of
’snapshot’ images, as the latter approach implicitly causes a training samples decreases. Nevertheless, the test accuracy
loss of information. remains on average higher than or approximately equal to
the state-of-the-art baselines in [23], [31] as the number of
TABLE VII
training data is reduced. up to the point of using only 20% of
C LASSIFICATION ACCURACY AND F1 SCORE OF THE PROPOSED
PIPELINE VS BASELINES IN THE LITERATURE FOR THE TASKS OF the training data (equivalent to 358 samples per activity).
ORIENTATION CLASSIFICATION , PC- BASED CLASSIFICATION , AND To conclude, in terms of different quantitative metrics
OVERALL HAR WITH 6 CLASSES . Diff INDICATES PERFORMANCE presented in this subsection, the proposed pipeline in general
DIFFERENCES COMPARED TO THE PROPOSED PIPELINE . has significantly superior performance over the baselines in 1)
Method Task Acc. Diff. F1 Diff. angle insensitivity, 2) classification accuracy, and 3) robustness
proposed HAR 87.1% 0% 86.7% 0% against a limited number of training samples.
baseline-2 HAR 74.9% -12.2% 73.6% -13.1%
baseline-3 HAR 70.1% -17.0% 68.9% -17.8%
baseline-4 HAR 65.3% -21.8% 64.5% -22.2%
baseline-5 HAR 62.7% -24.4% 61.9% -24.8%
baseline-6 HAR 58.4% -28.7% 58.2% -28.5% D. Results for Different SNR
proposed Ori. classification 97.5% 0% 97.3% 0%
baseline-2 Ori. classification 75.1% -22.4% 73.9% -23.4%
This section focuses on one of the most influential pa-
baseline-3 Ori. classification 53.4% -44.1% 52.8% -44.5% rameters in radar systems, SNR. Because of the randomness
baseline-4 Ori. classification 56.3% -41.2% 56.4% 40.9% nature of additive noise, data generation, training and testing
baseline-5 Ori. classification 62.8% -34.7% 61.5% -34.8%
baseline-6 Ori. classification 62.1% -35.4% 61.6% -35.7% are repeated for five independent realizations and averaged for
proposed PC classification 99.0% 0% 97.2% 0% every value of considered SNR, as described in Section III.
baseline-2 PC classification 86.0% -13.0% 85.8% -11.4% Figure 6 shows the average test accuracy and the standard
baseline-3 PC classification 87.5% -11.5% 87.4% -9.8%
baseline-4 PC classification 83.5% -15.5% 82.9% -14.3% deviation of these five realizations in terms of varying SNR
baseline-5 PC classification 80.2% -18.8% 79.9% -17.3% levels. As can be seen, the classification performance decrease
baseline-6 PC classification 81.8% -17.2% 81.4% -15.8% almost linearly along with decreasing SNR levels, and the
accuracy could drop to nearly 50% for an SNR of 8dB, which
Radar data measurement is generally more complex and is approximately 35% lower than the measured data (assumed
time-consuming than using vision-based sensors, thus limiting to be noise-free in this evaluation). These results suggest that
the typical radar-based HAR dataset to be much smaller than noisy data could significantly undermine the performance of
the computer vision datasets. To be more specific, Yang’s work the proposed pipeline. Last and most importantly, the perfor-
[31] and Kim’s work [23] included 60 and 288 samples per mance gain due to the proposed replacing of max pooling
activity, respectively, whereas our dataset includes roughly with average pooling is clearly shown by comparing the blue
300 training samples per activity. Therefore, an important curve with the red in Figure 6. This appears to indicate that
comparison is about the classification pipeline performance average pooling as a symmetric function fits better the task
with respect to a limited number of training samples. Figure of processing noisy radar data compared to the original max
5 presents the test accuracy of the baselines and the proposed pooling in [29].
pipeline trained with only a randomly selected percentage of For data-driven classification methods, it is also interesting
the training data available from the measurements. The results to evaluate whether differences in SNR between training
in Figure 5 show that the performance of the proposed pipeline and test data could influence classification performance. In
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
8 IEEE SENSORS JOURNAL, VOL. XX, NO. XX, XXXX 2017
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
AUTHOR et al.: IMAGING RADAR-BASED HUMAN ACTIVITY RECOGNITION 9
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/JSEN.2022.3175618, IEEE Sensors
Journal
10 IEEE SENSORS JOURNAL, VOL. XX, NO. XX, XXXX 2017
[11] M. Jia, S. Li, J. Le Kernec, S. Yang, F. Fioranelli, and O. Romain, [32] K. Simonyan and A. Zisserman, “Very deep convolutional networks for
“Human activity classification with radar signal processing and machine large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
learning,” in 2020 International Conference on UK-China Emerging [33] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image
Technologies (UCET). IEEE, 2020, pp. 1–5. recognition,” in Proceedings of the IEEE conference on computer vision
[12] S. Zhu, J. Xu, H. Guo, Q. Liu, S. Wu, and H. Wang, “Indoor human and pattern recognition, 2016, pp. 770–778.
activity recognition based on ambient radar with signal processing [34] G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, “Densely
and machine learning,” in 2018 IEEE international conference on connected convolutional networks,” in Proceedings of the IEEE confer-
communications (ICC). IEEE, 2018, pp. 1–6. ence on computer vision and pattern recognition, 2017, pp. 4700–4708.
[13] Y. Kim and H. Ling, “Human activity classification based on micro- [35] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez,
Doppler signatures using a support vector machine,” IEEE transactions Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in
on geoscience and remote sensing, vol. 47, no. 5, pp. 1328–1337, 2009. neural information processing systems, vol. 30, 2017.
[14] R. Zhang and S. Cao, “Real-time human motion behavior detection via
CNN using mmwave radar,” IEEE Sensors Letters, vol. 3, no. 2, pp.
1–4, 2018.
[15] B. Erol, S. Z. Gurbuz, and M. G. Amin, “Motion classification using
kinematically sifted acgan-synthesized radar micro-Doppler signatures,”
IEEE Transactions on Aerospace and Electronic Systems, vol. 56, no. 4, Yubin Zhao received the B.S. degree in elec-
pp. 3197–3213, 2020. trical engineering from University of Electronic
[16] L. Tang, Y. Jia, Y. Qian, S. Yi, and P. Yuan, “Human activity recognition Science and Technology of China, Chengdu,
based on mixed CNN with radar multi-spectrogram,” IEEE Sensors China, in 2019 and the M.S. degree in electrical
Journal, vol. 21, no. 22, pp. 25 950–25 962, 2021. engineering from Delft University of Technology,
[17] J. Zhu, H. Chen, and W. Ye, “A hybrid CNN–LSTM network for the Delft, The Netherlands, in 2022.
classification of human activities based on micro-Doppler radar,” IEEE
Access, vol. 8, pp. 24 713–24 720, 2020.
[18] H. Li, A. Shrestha, H. Heidari, J. Le Kernec, and F. Fioranelli, “Bi-
LSTM network for multimodal continuous human activity recognition
and fall detection,” IEEE Sensors Journal, vol. 20, no. 3, pp. 1191–1201,
2019.
[19] R. G. Guendel, M. Unterhorst, E. Gambi, F. Fioranelli, and A. Yarovoy,
“Continuous human activity recognition for arbitrary directions with
distributed radars,” in 2021 IEEE Radar Conference (RadarConf21). Alexander Yarovoy (F’15) graduated from the
IEEE, 2021, pp. 1–6. Kharkov State University, Ukraine, in 1984 with
[20] Y. Zhao, R. Gundel, A. Yarovoy, and F. Fioranelli, “Distributed radar- the Diploma with honor in radiophysics and
based human activity recognition using vision transformer and CNNs,” electronics. He received the Candidate Phys.
in presented at the 18th European Radar Conference, London, UK, 2022. & Math. Sci. and Doctor Phys. & Math. Sci.
[21] D. P. Fairchild and R. M. Narayanan, “Multistatic micro-Doppler radar degrees in radiophysics, in 1987 and 1994, re-
for determining target orientation and activity classification,” IEEE spectively. In 1987, he joined the Department
Transactions on Aerospace and Electronic Systems, vol. 52, no. 1, pp. of Radiophysics at the Kharkov State University
512–521, 2016. as a Researcher and became a Full Professor
[22] F. Fioranelli, M. Ritchie, and H. Griffiths, “Aspect angle depen- there, in 1997. From September 1994 through
dence and multistatic data fusion for micro-Doppler classification of 1996 he was with Technical University of Ilme-
armed/unarmed personnel,” IET Radar, Sonar & Navigation, vol. 9, nau, Ilmenau, Germany as a Visiting Researcher. Since 1999, he is with
no. 9, pp. 1231–1239, 2015. the Delft University of Technology, Delft, the Netherlands. Since 2009,
[23] Y. Kim, I. Alnujaim, and D. Oh, “Human activity classification based he leads there a Chair of Microwave Sensing, Systems and Signals.
on point clouds measured by millimeter wave MIMO radar with deep He has authored and coauthored more than 450 scientific or technical
recurrent neural networks,” IEEE Sensors Journal, vol. 21, no. 12, pp. papers, six patents and fourteen book chapters. His main research
13 522–13 529, 2021. interests are in high-resolution radar, microwave imaging and applied
[24] D. Nickalls, J. Wu, and N. Dahnoun, “A real-time and high performance electromagnetics (in particular, UWB antennas). He is the recipient of
posture estimation system based on millimeter-wave radar,” in 2021 10th the European Microwave Week Radar Award for the paper that Best
Mediterranean Conference on Embedded Computing (MECO). IEEE, advances the state-of-the-art in radar technology, in 2001 (together
2021, pp. 1–4. with L.P. Ligthart and P. van Genderen) and 2012 (together with T.
[25] J. Le Kernec, F. Fioranelli, C. Ding, H. Zhao, L. Sun, H. Hong, Savelyev). In 2010, together with D. Caratelli Prof. Yarovoy got the Best
J. Lorandel, and O. Romain, “Radar signal processing for sensing in Paper Award of the Applied Computational Electromagnetic Society
assisted living: The challenges associated with real-time implementation (ACES). He served as an Associated Editor of the International Journal
of emerging algorithms,” IEEE Signal Processing Magazine, vol. 36, of Microwave and Wireless Technologies from 2011 till 2018 and as a
no. 4, pp. 29–41, 2019. Guest Editor of five special issues of the IEEE Transactions and other
[26] S. Z. Gurbuz and M. G. Amin, “Radar-based human-motion recognition journals. He served as the General TPC chair of the 2020 European
with deep learning: Promising applications for indoor monitoring,” IEEE Microwave Week (EuMW’20), as the Chair and TPC chair of the 5th
Signal Processing Magazine, vol. 36, no. 4, pp. 16–28, 2019. European Radar Conference (EuRAD’08), as well as the Secretary of
[27] F. J. Abdu, Y. Zhang, and Z. Deng, “Activity classification based on fea- the 1st European Radar Conference (EuRAD’04). He served also as
ture fusion of FMCW radar human motion micro-Doppler signatures,” the co-chair and TPC chair of the Xth International Conference on GPR
IEEE Sensors Journal, vol. 22, no. 9, pp. 8648–8662, 2022. (GPR2004). In the period 2008 to 2017, he served as Director of the
[28] Q. An, S. Wang, A. Hoorfar, W. Zhang, H. Lv, S. Li, and J. Wang, European Microwave Association (EuMA).
“Range-max enhanced ultra-wideband micro-Doppler signatures of
behind wall indoor human activities,” 2020. [Online]. Available:
Francesco Fioranelli (Senior Member, IEEE)
https://arxiv.org/abs/2001.10449
received the Ph.D. degree with Durham Uni-
[29] C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on
versity, Durham, UK, in 2014. He is currently
point sets for 3D classification and segmentation,” in Proceedings of the
a tenured Assistant Professor at TU Delft, The
IEEE conference on computer vision and pattern recognition, 2017, pp.
Netherlands, and was an Assistant Professor
652–660.
with the University of Glasgow (2016–2019), and
[30] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification
a Research Associate at University College Lon-
with deep convolutional neural networks,” Advances in neural informa-
don (2014–2016). His research interests include
tion processing systems, vol. 25, 2012.
the development of radar systems and automatic
[31] Y. Yang, C. Hou, Y. Lang, T. Sakamoto, Y. He, and W. Xiang,
classification for human signatures analysis in
“Omnidirectional motion classification with monostatic radar system
healthcare and security, drones and UAVs de-
using micro-Doppler signatures,” IEEE Transactions on Geoscience and
tection and classification, automotive radar, wind farm, and sea clutter.
Remote Sensing, vol. 58, no. 5, pp. 3574–3587, 2019.
1530-437X (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: TU Delft Library. Downloaded on June 01,2022 at 06:34:10 UTC from IEEE Xplore. Restrictions apply.