SP Nov2022
SP Nov2022
In keeping with IEEE’s continued commitment to providing Your research will also be exposed to 5 million unique
options supporting the needs of all authors, in 2020, IEEE monthly users of the IEEE Xplore® Digital Library.
introduced the high-quality publication, the IEEE Open Journal The high-quality IEEE Open Journal of Signal Processing will
of Signal Processing. draw on IEEE’s expert technical community’s continued
In recognition of author funding difficulties during this commitment to publishing the most highly cited content.
unprecedented time, the IEEE Signal Processing Society, for The editor-in-chief is the distinguished Prof. Brendt
2022, is offering a reduced APC of USD$995 with no page Wohlberg, who specializes in signal and image processing,
limits. (This offer cannot be combined with any other inverse problems, and computational imaging.
discounts.) The rapid peer-reviewed process targets a publication
We invite you to have your article peer-reviewed and published time frame within 10-15 weeks for most accepted papers.
in the new journal. This is an exciting opportunity for your This journal is fully open and compliant with funder
research to benefit from the high visibility and interest the mandates, including Plan S.
journal will generate.
FEATURES 73 SP Forum
Scientific Integrity and Misconduct
in Publications
Luigi Longobardi, Tony VenGraitis,
and Christian Jutten
18 RETHINKING BAYESIAN
LEARNING FOR DATA ANALYSIS
Lei Cheng, Feng Yin, 76 Tips & Tricks
Sergios Theodoridis, Fast and Accurate Linear Fitting
Sotirios Chatzis, for an Incompletely Sampled Gaussian
and Tsung-Hui Chang Function With a Long Tail
Kai Wu, J. Andrew Zhang,
and Y. Jay Guo
53 RADIO MAP ESTIMATION
Daniel Romero and Seung-Jun Kim 85 Applications Corner
The SONICOM Project: Artificial
Intelligence-Driven Immersive Audio,
From Personalization to Modeling
Lorenzo Picinali, Brian FG Katz,
Michele Geronazzo, Piotr Majdak,
Arcadio Reyes-Lecuona,
and Alessandro Vinciarelli
ON THE COVER
A variety of topics are addressed in this issue including
scientific integrity, rethinking Bayesian learning and radio
map estimation.
COLUMNS
6 Special Reports
Signal Processing at the Epicenter
of Ground-Shaking Research
John Edwards
10 SP Everywhere
Artistic Text Style Transfer
Xinhao Wang, Shuai Yang,
PG. 10 Wenjing Wang, and Jiaying Liu PG. 85
IEEE SIGNAL PROCESSING MAGAZINE (ISSN 1053-5888) (ISPREG) is published bimonthly by the Institute of Electrical and Electronics Engineers, Inc., 3 Park Avenue, 17th Floor, New York,
NY 10016-5997 USA (+1 212 419 7900). Responsibility for the contents rests upon the authors and not the IEEE, the Society, or its members. Annual member subscriptions included in Society fee.
Nonmember subscriptions available upon request. Individual copies: IEEE Members US$20.00 (first copy only), nonmembers US$248 per copy. Copyright and Reprint Permissions: Abstracting
is permitted with credit to the source. Libraries are permitted to photocopy beyond the limits of U.S. Copyright Law for private use of patrons: 1) those post-1977 articles that carry a code at the bot-
tom of the first page, provided the per-copy fee is paid through the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923 USA; 2) pre-1978 articles without fee. Instructors are
permitted to photocopy isolated articles for noncommercial classroom use without fee. For all other copying, reprint, or republication permission, write to IEEE Service Center, 445 Hoes Lane,
Piscataway, NJ 08854 USA. Copyright © 2022 by the Institute of Electrical and Electronics Engineers, Inc. All rights reserved. Periodicals postage paid at New York, NY, and at additional mailing offices.
Postmaster: Send address changes to IEEE Signal Processing Magazine, IEEE, 445 Hoes Lane, Piscataway, NJ 08854 USA. Canadian GST #125634188 Printed in the U.S.A.
IEEEIEEE
SIGNAL PROCESSING
SIGNAL MAGAZINE| |November
MAGAZINE
PROCESSING 2022|
May 2021 | 1
IEEE Signal Processing Magazine
DEPARTMENTS EDITOR-IN-CHIEF
Christian Jutten—Université Grenoble Alpes,
ASSOCIATE EDITORS—COLUMNS AND FORUM
Ulisses Braga-Neto—Texas A&M University,
France USA
3 From the Editor Cagatay Candan—Middle East Technical
Scientific Integrity: A Duty for Researchers AREA EDITORS University, Turkey
Christian Jutten Feature Articles Wei Hu—Peking University, China
Laure Blanc-Féraud—Université Côte d’Azur, Andres Kwasinski—Rochester Institute of
4 President’s Message France Technology, USA
Starting the Ethics Discussion Xingyu Li—University of Alberta, Edmonton,
in Our Community Special Issues
Alberta, Canada
Athina Petropulu Xiaoxiang Zhu—German Aerospace Center,
Xin Liao—Hunan University, China
Germany
Piya Pal—University of California San Diego,
3 Dates Ahead
Cover Columns and Forum USA
Rodrigo Capobianco Guido—São Paulo Hemant Patil—Dhirubhai Ambani Institute
State University (UNESP), Brazil of Information and Communication
H. Vicky Zhao—Tsinghua University, Technology, India
R.P. China Christian Ritz—University of Wollongong,
e-Newsletter Australia
Behnaz Ghoraani—Florida Atlantic University,
USA ASSOCIATE EDITORS—e-NEWSLETTER
Social Media and Outreach Abhishek Appaji—College of Engineering, India
Emil Björnson—KTH Royal Institute of Technology, Subhro Das—MIT-IBM Watson AI Lab,
Sweden IBM Research, USA
Anubha Gupta—Indraprastha Institute of Information
EDITORIAL BOARD Technology Delhi (IIIT), India
Hamid Palangi—Microsoft Research Lab (AI), USA
Massoud Babaie-Zadeh—Sharif University of
Technology, Iran
Waheed U. Bajwa—Rutgers University,
IEEE SIGNAL PROCESSING SOCIETY
Athina Petropulu—President
USA
Min Wu—President-Elect
Caroline Chaux—French Center of National
Ana Isabel Pérez-Neira—Vice President,
Research, France
Conferences
Mark Coates—McGill University, Canada
Shrikanth Narayanan—VP Education
Laura Cottatellucci—Friedrich-Alexander
K.V.S. Hari—Vice President, Membership
University of Erlangen-Nuremberg, Germany
Marc Moonen—Vice President, Publications
Davide Dardari—University of Bologna, Italy
Alle-Jan van der Veen—Vice President, Technical
Mario Figueiredo—Instituto Superior Técnico,
Directions
University of Lisbon, Portugal
Sharon Gannot—Bar-Ilan University,
IEEE SIGNAL PROCESSING SOCIETY STAFF
Israel
William Colacchio—Senior Manager, Publications
Yifan Gong—Microsoft Corporation, USA
and Education Strategy and Services
Rémi Gribonval—Inria Lyon, France Rebecca Wollman—Publications Administrator
Joseph Guerci—Information Systems
Laboratories, Inc., USA
Ian Jermyn—Durham University, U.K.
COVER 3 Ulugbek S. Kamilov—Washington University, USA IEEE PERIODICALS MAGAZINES DEPARTMENT
Patrick Le Callet—University of Nantes, Sharon Turk, Journals Production Manager
©SHUTTERSTOCK.COM/G/SUNSINGER
SCOPE: IEEE Signal Processing Magazine publishes tutorial-style articles on signal processing research and
applications as well as columns and forums on issues of interest. Its coverage ranges from fundamental principles
to practical implementation, reflecting the multidimensional facets of interests and concerns of the community. Its
mission is to bring up-to-date, emerging, and active technical developments, issues, and events to the research,
IEEE prohibits discrimination, harassment, and bullying. educational, and professional communities. It is also the main Society communication platform addressing important
For more information, visit issues concerning all members.
http://www.ieee.org/web/aboutus/whatis/policies/p9-26.html.
E
thics in science is essential for vari- cess are also unacceptable and must be article. We also discussed corrective
ous reasons and is a duty for scien- met with punishment. actions, including a few new ones,
tists. The full sense of the word ethics In my view, scientific integrity is es- such as sending a letter informing the
may differ according to languages and sential and a duty for all scientists for employer of a scientist determined to
countries. For instance, in France, we the following reasons: be guilty of misconduct.
typically make a distinction between ■ We have a duty to Society, especially
ethics and scientific integrity, while since most of us are funded by pub- In this issue
scientific integrity is a part of ethics lic money, to conduct our research in Readers will again find a wide range
in the United States. For instance, the a manner that instills confidence and of interesting articles, with two feature
article “The Submerged Part of the AI- does not instead erode it. articles and a collection of five column
Ceberg,” [1] published in the September ■ Many scientists are involved in the articles. The feature article, “Rethink-
issue of IEEE Signal Processing Maga- training and supervision of young ing Bayesian Learning for Data Analy-
zine (SPM), discusses the ethical issues researchers. It is thus our duty to sis,” proposes a comprehensive tutorial
concerning the research in our domain, inculcate in them sound research on the Bayesian approach with priors
i.e., a set of philosophical reflections on practices that are based on scientif- promoting sparsity, which is illustrated
the interest and usefulness of our work ic integrity. on three typical tools used in data sci-
for humanity and for the Earth. These IEEE’s rules for good practices in pub- ence: deep neural networks, tensor de-
questions are complex and must prompt lications, reviews, and so on are detailed composition, and Gaussian processes.
debates between scientists and society. in the IEEE Publication Services and The second feature article, “Radio
Conversely, scientific integrity is a set Products Board Operations Manual Maps Estimation,” is a tutorial on radio
of good practices in the sciences, which [3]. For a simple, fast, and comprehen- map, which provide a radio-frequency
has not been discussed but is strictly sive overview of this manual, I posed spectrum landscape, with many appli-
applied. All of the learned scientific a few questions to Luigi Longobardi, cations in wireless communications
societies promote integrity. Of course, director of publishing ethics and con- and networking.
IEEE, as such and also as a publisher duct at IEEE, and Tony VenGraitis, Among the columns, you will find a
of many scientific journals, strongly program manager of publication eth- “Tips & Tricks,” which proposes an ef-
supports scientific integrity. Scien- ics for IEEE publications, and their ficient method for estimating an incom-
tific integrity and misconduct are at answers are detailed in “Scientific pletely sampled Gaussian function; an
the core of a few talks [2] at the an- Integrity and Misconduct in Publica- SP Forum,” which presents a review of
nual IEEE Panels of Editors meeting, tions” in this month’s issue of SPM methods for producing text with artis-
which bring together volunteers and on pages 73–75. tic effects; and a presentation of some
staff members to discuss assorted pub- Publishing-related misconduct has of the results from the European project
lication-related topics. Serious failures also been discussed in recent meetings SONICOM, which focuses on person-
related to scientific integrity include of the SPS Publication Board. Editors- alized immersive audio. Finally, John
plagiarism and data falsification or in-chief are tasked with confronting Edwards, as he has done for almost
fabrication, but misconduct related to misconduct and applying the appro- every SPM issue since 2011, offers an
authorship and during the review pro- priate corrective measure in response article in the “Special Reports” cat-
to it. A preventive step is the “custom egory. The article, “Signal Processing
questions” that must be completed in
Digital Object Identifier 10.1109/MSP.2022.3198298
Date of current version: 27 October 2022 ScholarOne when authors submit an (continued on page 84)
I
EEE members are bound by the IEEE deploy a product. It is only recently that When a technology promises obvi-
Code of Ethics [1]. By becoming IEEE in engineering schools located in the ous benefits, it may be easy to overlook
members, we commit ourselves to the United States, we started looking into potential harms to society and the envi-
highest ethical and professional conduct. ethical constraints, i.e., the impact of a ronment. Take, for example, artificial
We agree to uphold the practices of ethi- technology on the environment in terms intelligence (AI), offering tremendous
cal design and sustainable development; of waste produced; opportunities by
protect the privacy of others; and prompt- the power required enabling human-
ly disclose factors that might endanger for development Can companies that are like abilities in
the public or the environment. We strive or for operation just starting up invest in machines. Biasing
to improve the public understanding of and safety of users; ethical practices, or is this AI training data
the implications of technology. We also and in general, the only feasible for already or algorithms can
seek to avoid conflicts of interest; accept impact to the job lead to ethics vio-
established names?
honest criticism of technical work; and market and society lations. Google’s
treat all persons fairly independent of in general. In the i m a ge r e c og n i-
race, religion, gender, disability, age, United States, these considerations are tion system wrongly classified images
national origin, sexual orientation, gen- studied in the senior year of college, dur- of minorities [2]; the Apple Card, ad-
der identity, or gender expression. ing the capstone design project, when it ministered by Goldman Sachs, has
How many have actually read the may be too late for students to appreciate come under recent scrutiny for gender
IEEE Code of Ethics? Further, is the ex- the importance of those issues. bias [3], and software used to sentence
istence of a code of ethics sufficient to As researchers and practitioners, criminals was shown to be biased
avoid negative impacts to our lives from we continue on the same path. Most against minorities [2], [4]. Deepfakes,
the technologies we develop? do recognize the importance of eth- enabled by machine learning (ML) and
A recurring story nowadays is that a ics and ethical behavior, but when we AI, can generate deceptive visual and
technology is introduced; its popularity publish or put out a product, we do not audio content.
grows fast; it reaches a point where it really think about whether we may be Further, AI and ML have a huge
has permeated our lives; and then, soci- in violation of ethics principles. Of carbon footprint. Training AI models
ety realizes what kinds of problems that course, doing the check is not always requires a lot of data, which need to be
technology may create. When dealing straightforward. stored in data centers, and this takes a
with societal impact as an afterthought, For a business, where the ultimate lot of energy. To train the final version
it is often too late to introduce meaning- success is the financial profit, the re- of Megatron-LM (a neural network for
ful fixes. quirement to uphold ethical principles natural language applications), Nvidia ran
Starting from school, we have is often met with skepticism. Will the 512 V100 GPUs over nine days, consum-
learned to focus on the technical part of ethical product be measured in the ing 27,648 kWh. This is the cost of three
the idea and not on the problems the idea framework companies already use to average U.S. households in a year [5]. And
can lead to down the road. The techni- account for value? How much does this is just one algorithm. Data centers
cal details are sufficient to publish an “ethics” add to the cost of doing busi- use huge amounts of energy, consuming
article or file for a patent, and ultimately, ness? Can companies that are just start- 10–50 times the energy per floor space
ing up invest in ethical practices, or of a typical commercial office building.
is this only feasible for already estab- Several professional organizations
Digital Object Identifier 10.1109/MSP.2022.3198299
Date of current version: 27 October 2022 lished names? (for example, NeurIPS [6] and CVPR
Interested(in(learning(about(upcoming(SPM(issues(or(open(calls(for(papers?(
Want(to(know(what(the(SPM(community(is(up(to?(
Follow%us%on%twitter%(@IEEEspm)%and/or%join%our%LinkedIn%group%%
(www.linkedin.com/groups/8277416/)%to%stay%in%the%know%and%share%your%ideas!%
E
arthquakes have afflicted people the earthquake source to the seismom- Trugman explains that understand-
throughout history. Today, thanks to eter recording it—or the person expe- ing the spatial patterns of ground
advanced technology, more is known riencing it. motion in different frequency bands is
about earthquakes, and more can be done Trugman, working in collaboration necessary to develop adequate build-
to protect people against them. Signal with Brown University geophysicist ing design codes for structures posi-
processing is playing a key role as inves- Victor Tsai, was able to access one tioned nea r active faults. Cur rent
tigators examine ways to combat one of of the densest observational arrays of seismic hazard models assume a uni-
humanity’s most deadly foes. research seismome- form ground motion
ters ever created. The In reality, the shaking prediction model—
A new view result was the most entirely ignoring the
A project at the University of Nevada at highly detailed mea-
amplitude can vary rad iat ion pat ter n.
Reno is changing how researchers view surements of the spa- systemically depending The resea rchers
potential earthquake damage. Using tial pattern of ground on the direction of the are driving into the
data collected by one of the densest motions in different earthquake source to the physics of the prob-
seismic arrays ever deployed, research- frequency bands. seismometer recording lem. Specifically,
ers revealed that earthquakes emit their The analysis was it—or the person they’re looking for
strongest seismic shockwaves in four based on measure- what might cause the
opposing directions. ments collected by
experiencing it. observed radiation
Daniel Trugman, an assistant pro- a set of two dozen pattern to differ from
fessor in the university’s Nevada Seis- small earthquakes recorded by the classic theoretical earthquake models.
mological Laboratory, reports that Large-n Seismic Survey in Oklahoma When earthquakes strike, they
the study focused on what’s called (LASSO), an array of 1,829 seismic release a sudden burst of seismic energy
the earthquake radiation pattern, a sensors deployed for 28 days in 2016 to at multiple frequencies, but the actual
technical term for the spatial pattern monitor a remote corner of the state cov- ground shaking people feel ranges from
of ground motion. The effect, which ering 15 by 20 mi. about 1 to 20 Hz. The study revealed
creates a pattern resembling a four- The array helped the research- that low-frequency energy–approxi-
leaf clover, has been widely known ers understand the physics behind the mately 1–10 Hz–is emitted from the
and obser ved for many yea rs but observed radiation pattern as well as the fault in four directions. This observa-
never before measured in such precise limitations of current physical models. tion is critical since buildings are highly
detail (Figure 1). “This study is definitely not the end of vulnerable to low-frequency waves. The
It turns out that, when an earth- the story, but just a starting point for a four-leaf clover pattern wasn’t found for
quake occurs, the shaking amplitude is long-term project,” Trugman says. “The higher-frequency waves, which traveled
not the same in all directions. In reali- idea dates all the way back to my under- at equal strength in all directions.
ty, the shaking amplitude can vary sys- graduate honors thesis at Stanford on The study’s focus was to examine
temically depending on the direction of a tangentially related topic, but started how the radiation pattern varied as a
in earnest once I began a serendipitous function of frequency. “To do this, we
collaboration with Victor Tsai, who was bandpass filtered the data into a range of
Digital Object Identifier 10.1109/MSP.2022.3182932
Date of current version: 27 October 2022 interested in the same problem.” overlapping frequency bands spanning
FIGURE 1. The four-leaf clover radiation pattern, comparing observations with data. (Source: Daniel Trugman; used with permission.)
W
ord art, which is text rendered with such as the burning flames in Figure 1 artistic text style transfer. First, we for-
properly designed appealing artistic and exquisite decorations in Figure 2. mulate the task. Second, we investigate
effects, has been a popular form of Manually creating vivid text effects and classify state-of-the-art methods
art throughout human history. Artistic text requires lots of time and a series of com- into nondeep- and deep-based methods,
effects are of great aesthetic value and plicated operations: observing the target introduce their core ideas, and discuss
symbolic significance. Decorating with glyphs, designing appropriate artistic their strengths and weaknesses. Third,
appropriate effects not only makes text effects, warping the texture to match the we present several benchmark datasets
more attractive but also significantly character shapes, and so on. It consumes and evaluation methods. Finally, we
enhances the atmosphere of a scene. hours of time even for well-trained summarize the current challenges in
Thus, artistic text effects are widely used designers. To produce word art more this field and propose possible direc-
in publicity and advertising. Some text conveniently and efficiently, artistic text tions for future research.
effects are simple, such as colors and style transfer has been proposed recently
shadows, while some can be complex, to automatically render text with given Task formulation
artistic effects. Artistic text style transfer aims at
In this article, we provide a compre- automatically turning plain text into
Digital Object Identifier 10.1109/MSP.2022.3196763
Date of current version: 27 October 2022 hensive overview of current advances in fantastic artworks with given artistic
Source Text S Source Effects S′ Target Text T Result T ′ Source Effects S′ Target Text T Result T ′
Input Output Input Output
(a) (b)
FIGURE 1. An overview of application scenarios of text style transfer methods. (a) Transfer of text effects with supervision. (b) Joint transfer of text effects
and fonts without supervision. (c) Transfer of effects from arbitrary style images without supervision.
(a) (b)
(c) (d)
FIGURE 2. Some results of different text style transfer methods. (a) User-interactive effects transfer. (b) Decorative elements effects transfer.
(c) Shape-matching effects transfer. (d) Joint effects and fonts transfer.
S: Source Text
S ′: Source Effects
S S′ T T′ S′ T T′ T: Target Text
(a) (b) T ′: Result
Human Interaction
Stylization
S S′ T T′ S S′ T T′
(c) (d)
FIGURE 3. A brief introduction to nondeep text style transfer methods. (a) Supervised effects transfer. (b) Unsupervised effects transfer. (c) Supervised
effects transfer with human interaction. (d) Supervised dynamic effects transfer.
Sketchy S S S′
S′ T T′ T Shape-Matching T T′
(a) (b)
FIGURE 4. The two categories of deep-based text style transfer methods. (a) Multistyle-per-model methods. (b) Per-style-per-model methods.
Table 1. A summary of the benchmark datasets for the text effect transfer task.
from Peking University, Beijing, Chi- References Appl., vol. 12, no. 2, pp. 1–22, 2016, doi: 10.1145/2818709.
[1] Y. Wexler, E. Shechtman, and M. Irani, “Space-
na, in 2015 and 2020, respectively. He time completion of video,” IEEE Trans. Pattern
SP
Rethinking Bayesian
Learning for Data Analysis
The art of prior and inference in sparsity-aware modeling
S
parse modeling for signal processing and ma-
chine learning, in general, has been at the fo-
cus of scientific research for over two decades.
Among others, supervised sparsity-aware
learning (SAL) consists of two major paths paved
by 1) discriminative methods that establish direct
input–output mapping based on a regularized cost
function optimization and 2) generative methods that
learn the underlying distributions. The latter, more
widely known as Bayesian methods, enable uncer-
tainty evaluation with respect to the performed pre-
dictions. Furthermore, they can better exploit related
prior information and also, in principle, can natural-
ly introduce robustness into the model, due to their
unique capacity to marginalize out uncertainties re-
lated to the parameter estimates. Moreover, hyper-
parameters (tuning parameters) associated with the
adopted priors, which correspond to cost function
regularizers, can be learned via the training data and
not via costly cross-validation techniques, which is,
in general, the case with the discriminative methods.
To implement SAL, the crucial point lies in the
choice of the function regularizer for discriminative
methods and the choice of the prior distribution for
Bayesian learning. Over the past decade or so, due
to the intense research on deep learning, emphasis
has been put on discriminative techniques. Howev-
er, a comeback of Bayesian methods is taking place
that sheds new light on the design of deep neural
networks (DNNs), which also establish firm links
with Bayesian models, such as Gaussian processes
(GPs), and also inspire new paths for unsupervised learning, recent advances in incorporating sparsity-promoting priors
such as Bayesian tensor decomposition. The goal of this arti- into three highly popular data modeling/analysis tools, name-
cle is two-fold. First, it aims to review, in a unified way, some ly, DNNs, GPs, and tensor decomposition. Second, it reviews
their associated inference techniques from different aspects,
including evidence maximization via optimization and varia-
Digital Object Identifier 10.1109/MSP.2022.3198201
Date of current version: 27 October 2022 tional inference (VI) methods. Challenges, such as the small
Bayesian Methods
Regularized
Optimization
Parametric Nonparametric
Models Models
Bayesian Philosophy
FIGURE 1. The organization of this article and a road map for the readers.
■■ Prior: Then, we specify a prior on the unknown parameters The predicted value of y ) can be acquired via n T x ), and
i. For mathematical tractability, we adopt an i.i.d. Gaussian the posterior variance, b -1 + x T) R x ), quantifies the uncer-
distribution as the prior: tainty about this point prediction. Rather than providing a
L point prediction as in the discriminative methods, Bayesian
p M (i; h p) = % N (i l; 0, a l-1) (10) methods advocate averaging all possible predicted values
l =1
via marginalization and are thus more robust against errone-
where a l is the precision associated with each i l and ous parameter estimates.
h p = a _ 6a 1, a 2, f, a L@T represents the hyperparameters
associated with the prior. Bayesian nonlinear nonparametric model:
■■ Evidence: After substituting the prior (10) and the likelihood GP regression example
(9) into (5) and performing the integration, we can derive To improve the data representation power of Bayesian linear
the following Gaussian evidence: parametric models, a lot of effort has been invested in designing
nonlinear and nonparametric models. A direct nonlinear gener-
p M (D; h) = N ^ y; 0, b -1 I + X A -1 X T h (11) alization of (8) is given by
where the diagonal matrix A _ diag " a , and I denotes the y = f (x) + v (15)
identity matrix. Here, we have h = 6h Tp , b@ .
T
■■ Posterior: Inserting the prior (10), the likelihood (9), and the where, instead of the linearity of (8), we employ a nonlinear
evidence (5) into Bayes’ theorem (1), the posterior can be functional dependence f (x) and let v be the noise term, as be-
shown to be the Gaussian distribution fore. Moreover, the randomness associated with the weight pa-
rameters i in (8) is now embedded into the function f (x) itself,
p M (i ; D; h) = N (i; n, R) (12)
which is assumed to be a random process. That is, the outcome/
where realization of each random experiment is a function instead of a
T single value/vector. Thus, in this case, we have to deal with pri-
n = bRX y (13a)
ors related to nonlinear functions directly rather than indirectly,
R = (A + bX T X )-1.(13b) i.e., by specifying a family of nonlinear parametric functions and
placing priors over the associated weight parameters.
Once again, taking the preceding linear regression as a
concrete example, we further demonstrate the merits of GP model
Bayesian learning in general, as follows: In the following, we introduce one representative model that
■■ Merit 1—Parameter learning with uncertainty quantifica- adopts this rationale, namely, the GP model for nonlinear regres-
tion: Using Bayes’ theorem, the posterior in (12) not only sion. The GP models constitute a special family of random pro-
provides a point estimate n in (13a) for the unknown param- cesses, where the outcome of each experiment is a function or a
eters i but also provides a covariance matrix R in (13b) that sequence. For instance, in signal processing, this can be a con-
describes to which extent the posterior distribution is cen- tinuous-time signal f(t) as a function of time t or a discrete-time
tered around the point estimate n. In other words, it quanti- signal f(n) in terms of the sequence index n. In this article, we
fies our uncertainty about the parameter estimate, which treat the GP model as a data analysis tool whose input that acts as
cannot be naturally obtained in any discriminative method. the argument in f ($) is a vector; i.e., vector x = 6x 1, x 2, f, x L@T .
For the preceding example, we have it MAP = it MMSE because For clarity, we give the definition of GP as follows [1], [24]: a
the posterior distribution p M (i ; D; h) follows a unimodal random process, f (x), is called a GP if and only if for any finite
Gaussian distribution. Of course, frequentist methods can number of points, x 1, x 2, f, x N , the associated joint probability
also construct uncertainty region/confidence intervals by tak- density function (pdf), p ^ f (x 1), f (x 2), f, f (x N )h, is Gaussian.
ing a few extra steps once the parameter estimates have been A GP can be considered an infinitely long vector of joint-
obtained. However, the Bayesian method provides, in one ly Gaussian distributed random variables, so it can be fully
go, the posterior distribution of the model parameters, from described by its mean function and covariance function, defined
which both a point estimate as well as the uncertainty region as follows:
can be optimally derived via the learning optimization step.
■■ Merit 2—Robust prediction via marginalization: After m (x) _ E 6 f (x)@ (16)
substituting (12) and (9) tailored to new observations into cov (x, xl ) _ E 6^ f (x) - m (x)h^ f (xl ) - m (xl ) h@ .(17)
0
frequency parameters, s, and the time difference parameters, x.
In the “The Art of Prior: Sparsity-Aware Modeling for Three
Case Studies” section, we introduce some optimal kernel design
–5 methods that were first built based on the spectral density in the
–10 –5 0 5 10 frequency domain and then transformed back to the original in-
x put domain.
(b)
0
By applying some classic conditional Gaussian results (see,
–1
e.g., [1], [19], and [24]) we can derive the posterior distribution
from the joint distribution in (22) as –2
–3
p ^ y ) ; y h + N ^ y ); m
r , Vr h (23) –10 –5 0 5 10
x
where the posterior mean (vector) and the posterior covariance Observations Sample 1 Sample 3
(matrix) are, respectively, Mean Function Sample 2 95% CR
(b)
r = K ^ X ), X h6K ^ X, X h + b I@ y (24)
m -1 -1
0.25 1
0.2 0.8
0.15 0.6
p (θ1, θ2)
p (θ1, θ2)
0.1 0.4
0.05 0.2
0 0
5 5
5 5
0 0
θ1 0 θ1 0
θ2 θ2
–5 –5 –5 –5
(a) (b)
FIGURE 4. The joint probability distribution of the model parameters in 2D space. (a) The Laplacian distribution. (b) The Gaussian distribution. The heavy-
tail Laplacian distribution peaks sharply around zero and falls slowly along the axes, thus promoting sparse solutions in a probabilistic manner. On the
contrary, the Gaussian distribution decays more rapidly along both dimensions when compared to the Laplacian distribution.
0.5
0.02
0.4
0.015
p (θ1, θ2)
0.3
p (θ1, θ2)
0.01 0.2
0.005 0.1
0 0
5 5
5 5
0 0
θ1 0 θ1 0
θ2 θ2
–5 –5 –5
5 –5
(a) (b)
FIGURE 5. The representative GSM prior distributions in 2D space. (a) The Student’s t distribution and (b) the horseshoe distribution. It can be seen that
these two distributions show different heavy-tail profiles and are both sparsity promoting.
Stick-Breaking Construction
π1 = u1
π2 = u1u2
π3 = u1u2u3
πj = ul, j = 1, 2, . . .
l=1 {0, 1} K→∞
...
...
α uj πj zij L ...
...
...
(a) 1 0 (b)
FIGURE 6. The implementation of the IBP via the stick-breaking construction. (a) The beta–Bernoulli model. (b) The binary matrix Z.
y1f
f
w12
y2f f
w22
...
. . .
. . .
3
f
w32 y2f + 1 = g yif w f
i2
i =1
...
y3f
FIGURE 7. A deep fully connected NN. The (a) first layer, (b) fth layer, (c) ( f + 1) th layer, and (d) Fth layer.
FIGURE 8. The node-wise sparsity-aware modeling for DNNs using GSM priors.
Sparsity-aware modeling using IBP prior During training, posterior estimates of the respective prob-
The previous approach to imposing sparsity inherits a major abilities are obtained, which then allow for a naturally arising
drawback that is shared by all techniques for designing DNNs. component omission (link pruning) mechanism by introducing
That is, the number of nodes per layer has to be specified and a cutoff threshold x ; any link/weight with an inferred posteri-
preselected. Of course, one may say that we can choose a very or below this threshold value is deemed unnecessary and can
large number of nodes and then harness “sparsity” to prune the be safely omitted from computations. This inherent capability
network. However, if one overdoes it, he/she soon runs into prob- renders the considered approach a fully automatic data-driven
lems due to overparameterization. In contrast, we now turn our principled paradigm for SAL based on explicit inference of
attention to nonparametric techniques. We assume that the nodes component utility based on dedicated latent variables.
per layer are theoretically infinite (in practice, a large enough By utilizing the aforementioned construction, we can easily
number) and then use the IBP prior to enforce sparsity (Figure 9). incorporate the IBP mechanism in conventional ReLU-based
In line with what has been said while introducing the IBP networks and perform inference. However, the flexibility of
(the “Indian Buffet Process Prior” section), we multiply each the link-wise formulation allows us to go one step further. In
f
weight, i.e., w ij, with a corresponding auxiliary (hidden) binary recent works, the stick-breaking IBP prior has been employed in
f
random variable, z ij . The required priors for these variables, conjunction with a radically different, biologically inspired, and
f af a f+1
{{z ij} i = 1} j = 1, are generated via the IBP prior. In particular, competition-based activation, namely, the stochastic local win-
f f+1
we define a binary matrix Z f ! R a # a , with its (ij)th element ner takes all (LWTA) [2], [3], [4]. In the general LWTA context,
f
being z ij for the fth layer. Due to the sparsity-promoting nature neurons in a conventional hidden layer are replaced by LWTA
of the IBP prior, most elements in Z f tend to be zero, nulling blocks consisting of competing linear units. In other words, each
f f f+1
the corresponding weights in {{w ij} ai = 1} aj = 1, due to the involved node includes a set of linear (inner product) units. When pre-
multiplication. This leads to an alternative sparsity-promoting sented with an input, each unit in each block computes its acti-
modeling for DNNs [2], [3]. vation; the unit with the strongest activation is deemed to be the
... ...
... ...
i th Node ith Node
... ...
j th Node jth Node
When the f
f f Binary Variable Modeling { zij }
wij × zij
f Via IBP Prior
zij = 0
FIGURE 9. The link-wise sparsity-aware modeling for DNNs using the IBP prior.
Input Layer
IBP and LWTA IBP and LWTA
Output Layer
k (x, xl ) = / a i k i (x, xl ) (37)
Layer Layer i=1
1
ξ=1 1 where the weights, a i, i = 1, 2, f, and Q,
1
z11 =1 can either be set manually or be opti-
x1
mized. Each one of these subkernels can
1
z11 =1 be any one of the known kernels or any
ξ=0
function that admits the properties that
define a kernel; see, e.g., [19]. One may
consider constructing such a kernel ei-
ther in the original input domain or in the
frequency domain. The most straightfor-
ward way is to linearly combine a set of
ξ=0 elementary kernels, such as the SE kernel,
z1a1K = 0
xa1 rational quadratic kernel, periodic kernel,
and so on, with varying kernel hyperpa-
z1a1K = 0
K ξ=1 rameters in the original input domain; see,
K
e.g., [24], [25], and [34]. For high-dimen-
sional inputs, one can first detect pairwise
FIGURE 10. The LWTA and IBP-based architecture. Bold edges denote active (effective) connections
(with z ikf = 1). Nodes with bold contours denote winner units; i.e., they correspond to p = 1 (we do
interactions, using a fast and generic one
not use p kjf, to unclutter the notation). Rectangles denote LWTA blocks. For simplicity, each LWTA developed recently in [35], between the
block includes two (J = 2) competing linear units for k = 1, 2, f, K. inputs and for each interaction pair, adopt
model with trigonometric basis functions; namely, k (t, t l ; h p) = k (x; h p) = / a i exp 6- 2r 2 x 2 v i2@ cos (2rxn i)
i=1
(45)
Q
f (x) = / a i cos (2r~ i x) + b i sin (2r~ i x) (42)
i=1 where h p = [a 1, a 2, f, a Q, n 1, n 2, f, n Q, v 21, v 22, f, v Q2 ]T
denotes the hyperparameters of the SM kernel to be optimized
where {cos (2r~ i x), sin (2r~ i x)} constitute one pair of basis and x _ x - xl , owing to the stationary assumption. For accurate
functions parameterized in terms of the center frequencies ~ i, approximation, however, we need to choose a large Q, which po-
i ! {1, 2, f, Q}, and the random weights, a i and b i, are inde- tentially leads to an overparameterized model with many redun-
pendent and follow the same Gaussian distribution, N (0, v 20 /Q) dant localized Gaussian components. Besides, optimizing the
(it is noteworthy that ~ ! [0, 1/2) represents a normalized fre- frequency and variance parameters is numerically difficult as a
quency, namely, the physical frequency over the sampling fre- nonconvex problem and often incurs bad local minima.
quency). Under such assumptions, f(x) can be regarded as a GP, To remedy the aforementioned numerical issue, in [7],
according to the “Bayesian Nonlinear Nonparametric Model: it was proposed to fix the frequency and variance param-
GP Regression Example” section, and the corresponding covari- eters, n 1, n 2, f, n Q, v 12, and v 22, f, v Q2 , in the original SM
ance/kernel function can be easily derived as kernel to some known grids and focus solely on the weight
parameters, a 1, a 2, f, a Q . The resulting kernel is called
2 2 Q
the grid SM (GridSM) kernel. By fixing the frequency and
k (x, xl ) =
v0
Q
T
z (x) z (x l ) =
v0
Q
/ cos (2r~ i (x - xl )) (43) variance parameters, the preceding GridSM kernel can be
i=1
regarded as a linear multiple kernel with Q basis subkernels,
where the feature mapping vector z (x) contains all Q pairs of k i (x) _ exp 6- 2r 2 x 2 v i2@ cos (2rxn i), and i = 1, 2, f, Q. In [7],
trigonometric basis functions. it was shown that for sufficiently small variance, each subker-
Usually, we favor a large value of Q, well exceeding the nel matrix has a low rank smaller than N/2, namely, half of the
expected number of effective components. If the frequency data size. Therefore, it falls under the formulation in (38). The
points are randomly sampled from the underlying spectral den- corresponding weight parameters of such an overparameterized
sity, denoted by Su (~), then (43) is equivalent to the random Fou- kernel can be obtained effectively via optimizing the evidence
rier feature approximation of a stationary kernel function [39]. function in (38), and the solution turns out to be sparse, as dem-
However, in [36], the center frequencies are optimized through onstrated in the “The Art of Inference: Evidence Maximization
maximizing the evidence function. As claimed in the paper, such and Variational Approximation” section.
S(ω) S(ω)
140 250
120
200
100
Spectral Density
Spectral Density
150
80
60 100
40
50
20
ω ω
0 0
0 0.1 0.2 0.3 0.4 0.5 0 0.05 0.1 0.15 0.2 0.25 0.3
Normalized Frequency Normalized Frequency
(a) (b)
FIGURE 11. A comparison of the SM kernel and the original sparse spectrum kernel in (43) for approximating the underlying spectral density. Herein, the
SM kernel employs a mixture of Gaussian basis functions (see the blue curves), while the original sparse spectrum kernel employs a mixture of Dirac
deltas (see the red vertical lines). (a) Band-pass shape spectrum. (b) Low-pass shape spectrum.
3D X1 Xl XL 3D X1 Xl XL
Tensor Data First Rank-1 lth Rank-1 Tensor Lth Rank-1 Tensor Data First Rank-1 l th Rank-1 Tensor Lth Rank-1
X Tensor Tensor X Tensor Tensor
Denotes Zero
FIGURE 12. The sparsity-aware modeling for rank-1 tensors using GSM priors.
where b is the precision (the inverse of variance) of the Gauss- where E q (·) 6 · @ denotes the expectation with respect to the varia-
ian noise. Since it is unknown, a noninformative prior [e.g., the tional pdf q (·) . The inference framework under the MF assump-
Jeffery prior, p (b) ? 1/b] can be employed. To promote the tion is termed MF-VI.
low rankness, for the lth columns of all the factor matrices, a Whether the integration in the denominator (60) has a closed
GSM sparsity-promoting prior with latent variance variable g l form is determined by the functional forms of the likelihood and
has been adopted; see the detailed discussions in the “Sparsity- the priors. In particular, if they are conjugate pairs within the
Aware Modeling for Tensor Decompositions” section. Usually, exponential family of probability distributions (see the discus-
the hyperparameters h in the adopted GSM priors are prese- sions in, e.g., [1] and [19]), the optimal variational pdf in (60)
lected to make the prior noninformative and thus need no fur- will accept a closed-form expression. Fortunately, for the Bayes-
ther optimization. The unknown parameters i include the fac- ian tensor CPD adopting the Gaussian likelihood and the GSM
tor matrices {A (p)} Pp = 1, the latent variance variables {g l} lL= 1 of prior for the columns of the factor matrices, this condition is
the GSM priors, and the noise precision b. Under the evidence usually satisfied, which enables the derivation of closed-form
maximization framework, the inference problem can be formu- updates in recent advances [10], [11], [12], [13], [14], [15].
lated as (51) [the expression of the objective function in (51) is
quite lengthy (see, e.g., [14]) and thus is not included here] with Remark 4
unknown parameters, To facilitate the algorithm derivation, the MF-VI imposes a fac-
torization structure on q (i), which implies the statistical inde-
i _ {{A }p = 1, {g l} lL= 1,
( p) P
b}(57) pendence of the variables i k, given the observed dataset D. If
this is not the case, the MF approximation will lead to a mis-
and the joint pdf match when approaching the ground truth posteriors. In general,
the MF-VI tends to provide posterior approximations that are
p (D, i) _ p (YX, {{A (p)}Pp = 1, {g l} lL= 1, b}) (58) more compact compared to the true ones, which means that the
posterior estimates are usually “overconfident” [19]. To achieve
which can be computed by the product of the likelihood and the more accurate posterior estimation, there is a research trend to
priors. employ more advanced variational approximation techniques
Without imposing any constraint on the pdf q (i), the optimal than MF approximation. For example, recent tensor-aided
solution is just the posterior, i.e., q ) (i) = p M (i ; YX), whose Bayesian DNN research [49] utilizes the kernelized Stein dis-
computation using Bayes’ theorem will, however, encounter crepancy to derive the inference algorithm that can approximate
the intractable multiple integration challenge. To get over this the posterior better. The interested reader may refer to [59] for
difficulty, modern approximate inference techniques propose to some recent advances in variational approximation methods.
solve problem (51) by further constraining q (i) into a function- Some computational and theoretical difficulties that are com-
al family F; i.e., q (i) ! F. It is hoped that the family F is as monly encountered in Bayesian tensor decompositions are sum-
flexible as possible to allow accurate posterior estimates and, at marized as follows. First, due to the block coordinate descent
the same time, simple enough to enable tractable optimization nature of MF-VI [59], it is crucial to choose informative initial
algorithm designs. values to avoid poor local minima. On the other hand, the associ-
Among all the functional families, the mean field (MF) fam- ated computational complexity is cubic with respect to the ten-
ily is undoubtedly the most favorable one in recent Bayesian ten- sor rank (see, e.g., [10], [11], [14], and [51]), which is high if
sor research [10], [11], [12], [13], [14], [15]. It assumes that the the initial tensor rank is set to a large value. Finally, it was found
variational pdf q (i) = P Kk = 1 q (i k), where i is partitioned into that the algorithm performance significantly degrades in some
mutually disjoint nonempty subsets i k (i.e., , Kk = 1 i k = i, and challenging regimes, e.g., a low signal-to-noise ratio (SNR) and/
+ Kk = 1 i k = Q). In the context of the Bayesian tensor CPD, the or high rank (see, e.g., [10], [11], [14], and [51]). To overcome
MF assumption states that these difficulties, suggestions for real algorithm implementa-
P tions are provided in the “Unsupervised Learning via Bayesian
q (i) = % q (A (p)) q ({g l} lL= 1) q (b).(59) Tensor Decompositions” section.
p=1
The factorized structure in (59) inspires the idea of block mini- Inference algorithms for Bayesian DNNs
mization in the optimization theory. In particular, for the ELBO The step of inference (training) for Bayesian DNNs follows the
maximization problem (51), specified after fixing the varia- same backpropagation-type of philosophy as that of training
tional pdfs {q (i j)} j ! k, the resulting subproblem that optimizes their deterministic counterparts. There are, however, two notable
x1 ×w
1kj zN ...
(µ1 , yk1
x2 kj ζ
1kj )
×w
...
0, for j ≠ j0
i=1
...
)
, ζ Lkj ... ykJ
(µ Lkj
zN
xL × w Lk j
...
FIGURE 13. The kth block of a stochastic LWTA layer. Input x = [x 1, x 2, f, x L] is presented to each unit, j = 1, 2, f, J, in the block. Assume that the index
of the winner unit is j = j 0 . Then, the output of the block is a vector with a single nonzero value at index j 0 .
ECG: electrocardiogram.
The training data, D, are used for optimizing the hyperparameters of the learning model, while the test data, D ), are used for evaluating the prediction accuracy. The num-
bers given in the last two columns are the training sample size and test sample size, respectively.
6,000 6,000
5,000 5,000
Downlink Data Usage
4,000 4,000
3,000 3,000
2,000 2,000
1,000 1,000
0 0
–1,000 –1,000
0 100 200 300 400 500 600 700 0 100 200 300 400 500 600 700
Time Time
Training Data Prediction Training Data Prediction
Test Data Uncertainty Region Test Data Uncertainty Region
(a) (b)
6,000 6,000
5,000 5,000
Downlink Data Usage
4,000
4,000
3,000
3,000
2,000
2,000
1,000
0 1,000
–1,000 0
0 100 200 300 400 500 600 700 0 100 200 300 400 500 600 700
Time Time
Training Data Prediction Training Data Test Data Prediction
Test Data Uncertainty Region
(c) (d)
FIGURE 14. A comparison of 5G wireless traffic prediction performance obtained from different models. (a) A GSMGP model with Q = 500 fixed grids
whose e MAPE = 0.28. (b) A SMGP model with Q = 500 modes whose e MAPE = 0.42. (c) A standard GP with a hybrid of three elementary kernels whose
e MAPE = 0.3. (d) An LSTM model whose e MAPE = 1.12. The gray shaded areas represent the uncertainty region (computed as the posterior variances) of
the GP models.
50
Social Group Number Estimates
30 Automatically
Determine
The Bayesian learning algorithm the Social Group
20 suggests that four groups are Number
capable of data interpretation.
10
0
20 40 60 80 100
Iteration Number
Social Group Clustering
Government
Legal Affairs
Clustering Results
(Name List Was
Given in [9, Table 5])
Trading/Top
Pipeline
Executive
Bayesian
Tensor CPD
0.15
Crisis Breaks/Investigations
Temporal Profiles
Change
0.1 of CEO
Trading/
Enron E-mail Corpus Interpretable
Top Executive Bankruptcy
Temporal Profiles
0.05 Pipeline
0
10 20 30 40 50
Month
FIGURE 15. The Bayesian tensor CPD for social group clustering. CEO: chief executive officer.
(b)
(c)
(d)
(e)
(f)
FIGURE 16. The experimental results for visual comparison on image completion with 80% missing data. (a) The ground truth images. (b) The images
with missing values. The results from (c) Bayesian TTD, (d) TTC-TV, (e) TMAC-TT, and (f) STTO.
References [25] Y. Xu, F. Yin, W. Xu, J. Lin, and S. Cui, “Wireless traffic prediction with scal-
able Gaussian process: Framework, algorithms, and verification,” IEEE J. Sel. Areas
[1] S. Theodoridis, Machine Learning: A Bayesian and Optimization Perspective,
Commun., vol. 37, no. 6, pp. 1291–1306, Jun. 2019, doi: 10.1109/JSAC.
2nd ed. San Diego, CA, USA: Academic, 2020.
2019.2904330.
[2] K. Panousis, S. Chatzis, and S. Theodoridis, “Nonparametric Bayesian deep
[26] J. Lee, J. Sohl-Dickstein, J. Pennington, R. Novak, S. Schoenholz, and Y.
networks with local competition,” in Proc. Int. Conf. Mach. Learn. (ICML), 2019,
Bahri, “Deep neural networks as Gaussian processes,” in Proc. Int. Conf. Learn.
pp. 4980–4988.
Representations (ICLR), Vancouver, BC, Canada, 2018.
[3] K. Panousis, S. Chatzis, A. Alexos, and S. Theodoridis, “Local competition and
[27] Y. C. Eldar and G. Kutyniok, Compressed Sensing: Theory and Applications.
stochasticity for adversarial robustness in deep learning,” in Proc. Int. Conf. Artif.
Cambridge, U.K.: Cambridge Univ. Press, 2012.
Intell. Statist. (AISTAT), 2021, vol. 130, pp. 3862–3870.
[28] M. Elad, Sparse and Redundant Representations: From Theory to
[4] K. Panousis, S. Chatzis, and S. Theodoridis, “Stochastic local winner-takes-all
Applications in Signal and Image Processing. New York, NY, USA: Springer
networks enable profound adversarial robustness,” in Proc. Adv. Neural Inf.
Science & Business Media, 2010.
Process. Syst. (NeurIPS), 2021.
[29] D. F. Andrews and C. L. Mallows, “Scale mixtures of normal distributions,” J.
[5] C. Louizos, K. Ullrich, and M. Welling, “Bayesian compression for deep
Roy. Statist. Soc., B (Methodol.), vol. 36, no. 1, pp. 99–102, Sep. 1974, doi:
learning,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2017, pp. 3288–
10.1111/j.2517-6161.1974.tb00989.x.
3298.
[30] D. P. Wipf and B. D. Rao, “Sparse Bayesian learning for basis selection,” IEEE
[6] S. Ghosh, J. Yao, and F. Doshi-Velez, “Model selection in Bayesian neural net-
Trans. Signal Process., vol. 52, no. 8, pp. 2153–2164, Aug. 2004, doi: 10.1109/
works via horseshoe priors,” J. Mach. Learn. Res., vol. 20, no. 182, pp. 1–46, Oct.
TSP.2004.831016.
2019.
[31] T. L. Griffiths and Z. Ghahramani, “The Indian buffet process: An introduction
[7] F. Yin, L. Pan, T. Chen, S. Theodoridis, and Z.-Q. Luo, “Linear multiple low-
and review,” J. Mach. Learn. Res., vol. 12, no. 32, pp. 1185–1224, Jan. 2011.
rank kernel based stationary Gaussian processes regression for time series,” IEEE
[Online]. Available: http://jmlr.org/papers/v12/griffiths11a.html
Trans. Signal Process., vol. 68, pp. 5260–5275, Sep. 2020, doi: 10.1109/
TSP.2020.3023008. [32] D. J. MacKay, “Probable networks and plausible predictions — A review of
practical Bayesian methods for supervised neural networks,” Netw., Comput. Neural
[8] T. Paananen, J. Piironen, M. R. Andersen, and A. Vehtari, “Variable selection
Syst., vol. 6, no. 3, pp. 469–505, Feb. 1995, doi: 10.1088/0954-898X_6_3_011.
for Gaussian processes via sensitivity analysis of the posterior predictive distribu-
tion,” in Proc. Int. Conf. Artif. Intell. Statist. (AISTAT), 2019, pp. 1743–1752. [33] J. Bridle, “Training stochastic model recognition algorithms as networks can
lead to maximum mutual information estimation of parameters,” in Proc. Adv.
[9] H. Kim and Y. W. Teh, “Scaling up the automatic statistician: Scalable structure
Neural Inf. Process. Syst. (NeurIPS), 1990, vol. 2, pp. 211–217.
discovery using Gaussian processes,” in Proc. Int. Conf. Artif. Intell. Statist.
(AISTAT), Lanzarote, Spain, 2018, vol. 84, pp. 575–584. [34] D. Duvenaud, J. R. Lloyd, R. Grosse, J. B. Tenenbaum, and Z. Ghahramani,
“Structure discovery in nonparametric regression through compositional kernel
[10] L. Cheng, Z. Chen, Q. Shi, Y.-C. Wu, and S. Theodoridis, “Towards flexible
search,” in Proc. Int. Conf. Mach. Learn. (ICML), Atlanta, GA, USA, 2013, pp.
sparsity-aware modeling: Automatic tensor rank learning using the generalized
1166–1174.
hyperbolic prior,” IEEE Trans. Signal Process., vol. 70, no. 1, pp. 1834–1849, Apr.
2022, doi: 10.1109/TSP.2022.3164200. [35] T. Zhang, F. Yin, and Z.-Q. Luo, “Fast generic interaction detection for model
interpretability and compression,” in Proc. Int. Conf. Learn. Representations
[11] L. Cheng, X. Tong, S. Wang, Y.-C. Wu, and H. V. Poor, “Learning nonnegative
(ICLR), 2022, pp. 1–29.
factors from tensor data: Probabilistic modeling and inference algorithm,” IEEE
Trans. Signal Process., vol. 68, pp. 1792–1806, Feb. 2020, doi: 10.1109/ [36] M. Lázaro-Gredilla, J. Quiñonero Candela, C. E. Rasmussen, and A. R.
TSP.2020.2975353. Figueiras-Vidal, “Sparse spectrum Gaussian process regression,” J. Mach. Learn.
Res., vol. 11, pp. 1865–1881, Aug. 2010.
[12] Y. Zhou and Y.-M. Cheung, “Bayesian low-tubal-rank robust tensor factoriza-
tion with multi-rank determination,” IEEE Trans. Pattern Anal. Mach. Intell., vol. [37] A. Wilson and R. P. Adams, “Gaussian process kernels for pattern discovery
43, no. 1, pp. 62–76, Jan. 2021, doi: 10.1109/TPAMI.2019.2923240. and extrapolation,” in Proc. Int. Conf. Mach. Learn. (ICML), Atlanta, GA, USA,
2013, pp. 1067–1075.
[13] L. Cheng, Y.-C. Wu, and H. V. Poor, “Probabilistic tensor canonical polyadic
decomposition with orthogonal factors,” IEEE Trans. Signal Process., vol. 65, no. [38] J. Shawe-Taylor and N. Cristianini, Kernel Methods for Pattern Analysis.
3, pp. 663–676, Feb. 2017, doi: 10.1109/TSP.2016.2603969. Cambridge, U.K.: Cambridge Univ. Press, 2004.
[14] Q. Zhao, L. Zhang, and A. Cichocki, “Bayesian CP factorization of incomplete [39] A. Rahimi and B. Recht, “Random features for large-scale kernel machines,” in Proc.
tensors with automatic rank determination,” IEEE Trans. Pattern Anal. Mach. Int. Conf. Neural Inf. Process. Syst., Vancouver, BC, Canada, 2007, pp. 1177–1184.
Intell., vol. 37, no. 9, pp. 1751–1763, Jan. 2015, doi: 10.1109/TPAMI.2015. [40] M. W. Seeger, “Bayesian inference and optimal design for the sparse linear
2392756. model,” J. Mach. Learn. Res., vol. 9, pp. 759–813, Jun. 2008.
[15] Z. Zhang and C. Hawkins, “Variational Bayesian inference for robust stream- [41] M. E. Tipping and A. Faul, “Fast marginal likelihood maximisation for sparse
ing tensor factorization and completion,” in Proc. IEEE Int. Conf. Data Mining Bayesian models,” in Proc. Int. Conf. Artif. Intell. Statist. (AISTAT), Key West,
(ICDM), Singapore, 2018, pp. 1446–1451, doi: 10.1109/ICDM.2018.00200. FL, USA, 2003, pp. 3–6.
[16] Y. Dai, T. Zhang, Z. Lin, F. Yin, S. Theodoridis, and S. Cui, “An interpretable [42] N. I. Achieser, Theory of Approximation. New York, NY, USA: Dover,
and sample efficient deep kernel for Gaussian process,” in Proc. Int. Conf. 1992.
Uncertainty Artif. Intell. (UAI), 2020, pp. 759–768.
[43] N. D. Sidiropoulos, L. De Lathauwer, X. Fu, K. Huang, E. E. Papalexakis, and C.
[17] L. Cheng, Y.-C. Wu, and H. V. Poor, “Scaling probabilistic tensor canonical Faloutsos, “Tensor decomposition for signal processing and machine learning,” IEEE
polyadic decomposition to massive data,” IEEE Trans. Signal Process., vol. 66, no. Trans. Signal Process., vol. 65, no. 13, pp. 3551–3582, Jul. 2017, doi: 10.1109/
21, pp. 5534–5548, Nov. 2018, doi: 10.1109/TSP.2018.2865407. TSP.2017.2690524.
[18] L. Cheng and Q. Shi, “Towards overfitting avoidance: Tuning-free tensor-aided [44] E. E. Papalexakis, N. D. Sidiropoulos, and R. Bro, “From k-means to higher-
multi-user channel estimation for 3D massive MIMO communications,” IEEE J. way co-clustering: Multilinear decomposition with sparse latent factors,” IEEE
Sel. Topics Signal Process., vol. 15, no. 3, pp. 832–846, Apr. 2021, doi: 10.1109/ Trans. Signal Process., vol. 61, no. 2, pp. 493–506, Jan. 2013, doi: 10.1109/
JSTSP.2021.3058019. TSP.2012.2225052.
[19] C. M. Bishop, Pattern Recognition and Machine Learning. New York, NY, [45] H. Becker, L. Albera, P. Comon, R. Gribonval, F. Wendling, and I. Merlet,
USA: Springer-Verlag, 2006. “Brain-source imaging: From sparse to tensor models,” IEEE Signal Process. Mag.,
[20] S. F. Gull, “Bayesian inductive inference and maximum entropy,” in vol. 32, no. 6, pp. 100–112, Nov. 2015, doi: 10.1109/MSP.2015.2413711.
Maximum-entropy and Bayesian Methods in Science and Engineering, G. J. [46] C. Chatzichristos, E. Kofidis, M. Morante, and S. Theodoridis, “Blind fMRI
Erickson and C. R. Smith, Eds. Dordrecht, The Netherlands: Springer-Verlag, 1988, source unmixing via higher-order tensor decompositions,” J. Neurosci. Methods,
pp. 53–74. vol. 315, pp. 17–47, Mar. 2019, doi: 10.1016/j.jneumeth.2018.12.007.
[21] G. Schwarz, “Estimating the dimension of a model,” Ann. Statist., vol. 6, no. 2, [47] H. Chen, F. Ahmad, S. Vorobyov, and F. Porikli, “Tensor decompositions
pp. 461–464, Jul. 1978, doi: 10.1214/aos/1176344136. in wireless communications and MIMO radar,” IEEE J. Sel. Topics Signal
[22] D. J. MacKay, “Bayesian interpolation,” Neural Comput., vol. 4, no. 3, pp. Process., vol. 15, no. 3, pp. 438–453, Apr. 2021, doi: 10.1109/JSTSP.2021.
415–447, May 1992, doi: 10.1162/neco.1992.4.3.415. 3061937.
SHUTTERSTOCK.COM/G/SKILLUP
R
adio maps characterize quantities of interest in radio com- Introduction
munication environments, such as the received signal Spectrum cartography comprises a collection of techniques
strength and channel attenuation, at every point of a geo- used to construct and maintain radio maps, which provide use-
graphical region. Radio map estimation (RME) typically ful information on the radio-frequency (RF) landscape, such
entails interpolative inference based on spatially distributed mea- as the received signal power, interference power, power spec-
surements. In this tutorial article, after presenting some representa- tral density (PSD), electromagnetic absorption, and channel
tive applications of radio maps, the most prominent RME methods gain across a geographic area; see, e.g., [1], [2], and [3]. A
are discussed. Starting from simple regression, the exposition quick overview of the most representative types of radio maps
gradually delves into more sophisticated algorithms, eventually is provided in Table 1.
touching upon state-of-the-art techniques. To gain insight into this Radio maps find a myriad of applications in wireless com-
versatile toolkit, illustrative toy examples will also be presented. munications and networking, such as network planning, inter-
ference coordination and mitigation, power control, resource
allocation, handoff management, multihop routing, dynamic
Digital Object Identifier 10.1109/MSP.2022.3200175
Date of current version: 27 October 2022 spectrum access, and cognitive radio networking tasks; see [4]
0.1
0 20 40 60 80 100
x(m)
Power map Unveil regions of high inter- The mobile user reports
Power (dBm)
–42
ference power measurements.
–44 Determine appropriate loca-
–46
tions for new base stations
0 20 40 60 80 100
x(m)
PSD Map 2,460 Maximize frequency reuse The mobile user reports
Frequency (MHz)
0 25 50 75
x (m)
Channel gain 0 Resource allocation for The user at position x1 • the environment changes.
map device-to-device communi- sends a pilot sequence.
cations The user at x2 sends an
estimate of the received
x 1(m)
0 50
x 2(m)
Although radio maps find applications in many domains, this table exemplifies their applicability in cellular communications for specificity. The x-coordinate indexes a point on a
road or railway.
"
600 LS Map Estimate (p)
m 1 = a 1 } 1 (x 1) + g + a S } S (x 1)
Measurements (mn)
Power (µW)
h 400
m S = a 1 } 1 (x S) + g + a S } S (x S). (2)
200
In practice, however, the measurements are noisy, and one may
use more than S of them to estimate the coefficients. Upon 0
defining a|= 6a 1, f, a S@<, m|= 6m 1, f, m N@<, ^W hn, s|= }s (x n),
and z|= 6z 1, f, z N@<, (2) can be extended to the case with
0 20 40 60 80 100
x (m)
N 2 S measurements as m = Wa + z. The least-squares (LS)
estimate of a is, therefore, at = argmin a < m - Wa <2 . Because FIGURE 1. An example of map estimation in 1D using a parametric estima-
the number S of parameters to be estimated does not depend on tor that knows the transmitter locations. The estimate is reasonably
the number N of measurements, this approach is termed para- accurate despite the low number of measurements.
Table 2. A comparison of the power map estimation methods discussed in this tutorial.
Method Input (Besides Measurements) Strengths Limitations
Linear parametric • Transmitter locations x , f, x
TX
1
TX
S • Simplicity • Inaccurate in non-LOS conditions
RME • Path loss law: e.g., • Closed form • Requires transmitter locations
} s (x) := 1/ < x - x TX
s <
2
• Accuracy in LOS conditions
• Can easily accommodate knowledge of trans-
mit antenna patterns
Kernel-based • Reproducing kernel l (x, xl ) • High flexibility • Sensitive to the choice of the kernel
learning • Loss L • Does not require transmitter locations • Depending on L, may require a
• Regularization parameter m numerical solver
• m must be tuned, e.g., via cross
validation
Kriging • Mean n p (x) and covariance • LMMSE optimality • Accurate covariance structure may be
Cov [p (x), p (xl )] of p • Closed form hard to obtain
• Measurement noise variance v 2z • Naturally suited to the customary log-normal • Requires user locations
shadowing model
• Estimation error that can be quantified
Sparsity-based • Discrete grid • Efficient algorithms available for obtaining • Prior knowledge on propagation char-
methods • Regularization parameter m a solution acteristics needed
• Recovered sparse solution readily interpretable • Errors due to grid mismatch
Matrix completion • Regular grid • Agnostic to propagation characteristics • Critical low-rank condition
• Regularization parameter m • Spatial correlation structure exploited • Sufficient number of measurements
required for stable interpolation
Dictionary • Dictionary size Q • Powerful union-of-subspace prior for spatial • Nonconvex optimization
learning • Regularization parameters m s patterns • Hyperparameter tuning necessary
and m L • Can accommodate rapid temporal dynamics
Deep learning • Terrain maps • Can learn propagation patterns from a dataset • Large amount of data required
• Vegetation maps • More accurate than other methods if sufficient • Training that is computationally
• Building height maps data are available [18] intensive
• Network architecture
• Training parameters
• Others
FIGURE 2. An example of map estimation by fitting a polynomial of degree where m 2 0 is a predetermined regularization parameter, and
13 via LS. The estimate is clearly unsatisfactory despite the fact that the
L is a loss function quantifying the deviation between the
estimate accurately fits most of the measurements.
observations " m n ,nN= 1 and the predictions " g (x n) ,nN= 1 pro-
duced by a candidate g. If the square loss L (m n, g (x n)) =
(m n - g (x n))2 is adopted, (5) becomes kernel ridge regression
800 α1κ(x, x1) α2κ(x, x2) (KRR) [19, Ch. 4] (Figures 4 and 5).
α3κ(x, x3) α4κ(x, x4) The RKHS norm of g (x) = R i3= 1 a i l ^ x, xli h is given by
600
α5κ(x, x5) Σ αiκ(x, xi)
Power (µW)
i
/ / a i a j l^ xli, xljh .(6)
3 3
400 < g < G|=
i =1 j =1
200
The term < g <2G in (5) can be replaced by other increasing
0 functions of < g < G, but this explanation considers just the spe-
cial case of < g <2G for simplicity.
0 20 40 60 80 100
x (m) To understand the role of the regularization term m < g < 2G
in (5), first note that L is typically designed so that its mini-
FIGURE 3. An example of a function in an RKHS obtained with the expan- mum is attained when g (x n) = m n . Thus, in the absence of the
sion in (4) with only five terms. regularization term, owing to the infinite degrees of freedom
800 800
True Map (p) True Map (p)
"
"
Power (µW)
400 400
200 200
0 0
0 20 40 60 80 100 0 20 40 60 80 100
x (m) x (m)
FIGURE 4. An example of a KRR estimate. As expected, the quality of the FIGURE 5. An example of a KRR estimate with more measurements than
fit is higher in regions with higher measurement density. in Figure 4. The fit is considerably better.
Terrain Map
Building Map
p (x)
"
Distance to DNN
Source
Distance
From Sensor
y (m)
produced by two sources radiating with a fixed height and
power in free space. A dataset can be generated where 40
each map is obtained by placing the sources at random 20
locations on the horizontal plane. Each map is, therefore, 0
uniquely identified by the four scalars corresponding to the 0 50 100
x (m)
locations of the sources. If the maps are defined on a
32 # 32 grid, they comprise 32 2 = 1, 024 points, which S = {1, 2} S = {1, 3}
100 100
means that these maps lie on a manifold of dimension four
80 80
embedded in a space of dimension 1,024.
This observation is corroborated in [18] by training an 60 60
y (m)
y (m)
y (m)
Terrain Map
p (x)
"
Sampling Mask
CNN
Measurements
Building Map
FIGURE 7. A global DNN estimator, which provides pt (x) for all values of on a grid.
–70 5
80 80 80
4
Power (dBm)
–80 60 60 60
Uncertainty
3
y (m)
–90
40 40 40 2
–100 1
20 20 20
0
0 0 0
0 20 40 60 80 0 20 40 60 80 0 20 40 60 80
x (m) x (m) x (m)
(a) (b) (c)
FIGURE 10. An example of a surveying operation with an autonomous UAV in an urban environment seen from above: the (a) true power, (b) estimated
power, and (c) uncertainty metric. White boxes denote buildings. Red and white crosses denote measurement locations.
30
sured power will depend on the sensor orientation. For this
20 reason, it may be convenient to estimate the angular spectrum
map p (x, i), which provides the angular power density
y (m)
10
received by a sensor at location x from direction i. Here, i
0 parameterizes the direction through, e.g., the azimuth and ele-
vation angles.
–10
0 20 40 If C (i - il ) denotes the antenna gain along direction i for a
x (m) sensor with orientation il , it follows that the power received by
(b) such a sensor when placed at x will be # C (i - il ) p (x, i) di.
If the sensor orientations associated with all measurements
FIGURE 11. (a) Free space. (b) The scenario with four walls. The color of are known, then each measurement is a noisy linear observa-
each point indicates the x-coordinate of the location estimate obtained by tion of p (x, i) and, therefore, the latter can be estimated. The
a sensor at that location. The black circles indicate the positions of the techniques described earlier for PSD map estimation in the
transmitters. In (a), the estimate accurately matches the true coordinate
when there is no multipath; thus, this image serves as a color bar. On
frequency domain can be adapted to this end, possibly upon
the other hand, in (b), the estimation error is large in the presence of discretizing the aforementioned integral. Specifically, p (x, i)
multipath. (Source: [45]; used with permission.) can be estimated for a discrete set of angle bins separately or
Γn,1 ( f )
Measure
Quantize
Power
f
Γn,2 ( f )
Measure
Quantize
Power
f
Down-
Conversion Concatenate To FC
Γn,L ( f )
Measure
Quantize
Power
f
FIGURE 12. To reduce the rate necessary to report measurements, sensors may use a bank of random filters. The energy of each filter is measured,
quantized, and sent to an FC that performs RME.
E
thics and scientific integrity are key ■■ During the COVID-19 pandemic, elaborated collectively by these two
concepts in scientific research. many results based on bad practices, persons of the IEEE Publishing Ethics
Ethics in science concerns e.g., rejection of patients with risks, Team (PET).
reflections on what is morally good voluntary wrong interpretation of SPM: Let us begin the discussion
or bad and on the values that motivate results, and so on, have been pub- with the topic of well-known and serious
our actions and their consequences. It lished in journals and largely distrib- scientific misconduct, which includes
appeals to our sense of morality and uted in social networks: these are data fabrication, data falsification,
responsibility toward society, humans, examples of data falsification, a typ- and plagiarism. Can you comment on
and, more generally, living beings. ical and serious misconduct in terms these three areas of misconduct?
Scientific integrity concerns the right of scientific integrity. PET: Misconduct takes many forms
way of conducting research practices. IEEE Societies, particularly the and may involve the actions of authors,
It is mandatory for society to enhance IEEE Signal Processing Society, are editors, reviewers, and publishers. As
confidence in scientists and in the sci- strongly involved in the respect of sci- you pointed out, the U.S. Office of
ences. It is also very important when entific integrity. Details about the main Research Integrity defines research mis-
scientists serve as mentors for young rules and good practices for authors as conduct as the “fabrication, falsification,
researchers and, thus, are responsible well as reviewers, associate editors, and or plagiarism in proposing, performing,
for teaching good practices and setting editors-in-chief (EICs) are detailed in or reviewing research, or in reporting
examples of integrity for their students the IEEE Publication Services and Prod- research results.” All of these consti-
and collaborators. ucts Board (PSPB) Operations Manual, tute conduct or behavior that violates or
These two concepts, ethics and scien- which is available at https://pspb.ieee. compromises the ethical standards and
tific integrity, are sometimes confused, org/images/files/files/opsmanual.pdf. expectations determined by the scien-
although they are very dissimilar. In That 138-page-long manual is tific publishing community. It is impor-
short, we can claim that ethics is related very complete extension. Nevertheless, tant that all parties are aware of what is
to philosophical reflections capable of most past and current IEEE Members expected of them and are educated and
opening discussions, whereas scientific have never read it. In this forum article, informed regarding industry best prac-
integrity is related to a set of good prac- we propose to establish a discussion tices, guidelines, and codes of conduct.
tices that must be rigorously and always about different facts of scientific and When authors submit articles to
applied, without discussion. publishing misconduct to discover rules IEEE, they are expected to comply
As examples, we have the following. of scientific integrity and what you can with IEEE policy regarding authorship
■■ The relevance of experiments on liv- do when you face such misconduct. For responsibilities. Among these responsi-
ing beings and that of implementing this purpose, IEEE Signal Processing bilities are the needs to ensure that data
invasive brain–computer interfaces Magazine (SPM) EIC Christian Jutten are accurate and free from inappropri-
for augmenting the capacities of interviewed Luigi Longobardi, direc- ate influence and that any republished
well-being for people are considered tor of Publishing Ethics and Conduct at content has been properly referenced
to be ethical issues. IEEE, and Tony VenGraitis, program and cited. While it is not always possible
manager of Publication Ethics, IEEE for editors and peer reviewers to evalu-
Publications, and they answered some ate the validity of data, it is still impor-
Digital Object Identifier 10.1109/MSP.2022.3196764
Date of current version: 27 October 2022 questions. The answers have been tant to examine the article’s premise and
F
itting experiment data onto a curve is a been applied in transferring the Gaussian of the iterative WLS substantially im-
common signal processing technique fitting into a linear fitting [4]. However, proved, but its accuracy is also greatly
to extract data features and establish the problem of the logarithmic transfor- enhanced, particularly for noisy and in-
the relationship between variables. Of- mation is that it makes the noise power completely sampled Gaussian functions
ten, we expect the curve to comply with vary over data samples, which can result with long tails. These will be demon-
some analytical function and then turn in biased Gaussian fitting. The weighted strated by simulation results.
data fitting into estimating the unknown least square (WLS) fitting is known to be
parameters of a function. Among analyti- effective in handling uneven noise back- Prior art and motivation
cal functions for data fitting, the Gaussian grounds [7]. However, as unveiled in [5], Let us start by elaborating on the sig-
function is the most widely used one due the ideal weighting for linear Gaussian nal model. A Gaussian function can be
to its extensive applications in numerous fitting is directly related to the unknown written as
science and engineering fields. To name Gaussian function. To this end, an itera-
(x - n) 2
just a few, the Gaussian function is highly tive WLS is developed in [5], starting f (x) = Ae
-
2v 2 , (1)
popular in statistical signal processing with using the data samples (which are
and analysis, thanks to the central limit noisy values of a Gaussian function) as where x is the function variable, and
theorem [1], and the Gaussian function weights and then iteratively reconstruct- A, n, and v are the parameters to be
frequently appears in the quantum har- ing the weights using the previously esti- estimated. They represent the height,
monic oscillator, quantum field theory, mated function parameters. location, and width of the function, re-
optics, lasers, and many other theories For the iterative WLS, the number spectively. Directly fitting f(x) can be
and models in physics [2]; moreover, the of iterations required for a satisfactory cumbersome due to the exponential
Gaussian function is widely applied in fitting performance can be large, par- function. A well-known opponent of
chemistry for depicting molecular orbit- ticularly when an incompletely sampled the exponential is the natural logarithm.
als, in computer science for imaging pro- Gaussian function with a long tail is Indeed, by taking the natural logarithm
cessing, and in artificial intelligence for given [see Figure 1(a) for such a case]. of both sides of (1), we can obtain the
defining neural networks. Establishing a good initialization is a following polynomial after some basic
Fitting a Gaussian function, or, sim- common strategy for improving the rearrangements:
ply, Gaussian fitting, is consistently of convergence speed and performance
ln ^ f (x)h = a + bx + cx 2, (2)
high interest to the signal processing of an iterative algorithm. Noticing the
community [3]–[6]. Since the Gaussian unavailability of a proper initialization where the coefficients a, b, and c are
function is underlain by an exponential for the iterative WLS, we aim to fill the related to the Gaussian function param-
function, it is nonlinear and not easy to blank by developing a high-quality one eters n, v, and A. Based on (1) and (2),
be fitted directly. One effective way of in this article. To do so, we introduce a it is easy to obtain
counteracting its exponential nature is to few signal processing tricks to develop
n = - b ; v = -1 ; A = e a - b /(4c) .
2
1 1
f (x)
f (x)
0.5 0.5
Tail Region
0 0
0 5 9 10 0 5 10
x x
(a) (b)
Principal
1 1
Region
f (x)
f (x)
0.5 0.5
0 0
0 5 6 10 0 5 10
x x
(c) (d)
FIGURE 1. (a) The noisy samples of the Gaussian function with A = 1, n = 9, and v = 1.3 are plotted, where the dashed curve is the function without
noise. (b) The iterative WLS, as illustrated in (9), is run 10 times, each time with 12 iterations and independently generated noise, leading to the fitting
results shown. (c) and (d) Other than resetting n = 6 , the same results are plotted as in (a) and (b), respectively. As shown, the range of x is [0, 10],
where the sampling interval is set as d x = 0.01.
ure 1(a) for an illustration. The tail re- mation approximates the integral of f(x) 2
#0 f (x) dx
t is esti-
gion is about half the whole sampled described earlier. Moreover, A +1 #
Nd x
f (x) dx
2 n - (Nd x - n)
region. Perform 12 numbers of itera- mated by
erf c m
tions based on (9) for 10 trials, each (e) 2r Av n
6A
t , nt @ = max y [n] . (12) =
adding independently generated noise 2 v 2
n
onto the same Gaussian function with Nd x - n
erf c m,
2r A v
+
A = 1, n = 9, and v = 1.3. The fitting Similar to how max ($) works in MAT- 2 v 2
results are given in Figure 1(b). We see LAB [8], nt is the index where the maxi- (13)
CRLB " vt b ,
/ f [n] 2 ( n - d x n) 4 y [0], y [1], f, y [N - 1] . That is, taking
*
also obtain an estimate of A, i.e.,
n = nt t = t in (22) leads to an asymptotical-
CRLB " vt a ,
= .
nt - 1
t = y 8nt + 8 L BB . (27)
/ f [n] 2 ( n - d x n) 4 ly optimal unbiased v estimation. A
2
n=0 Note that f [n] used for calculating
(21) *
t is unavailable. Thus, we replace f [n] Use these two estimates obtained in
An insight from this result is that we with its noisy version y [n], attaining the the proposed v estimators, as given in
only need a linear combination of the following practically usable coefficient: (18). Then, combine the two estimates,
two v estimates obtained in (18) as done in (22), with the optimal com-
N-1
to achieve an asymptotically optimal bining coefficient given in (25). This
combined estimation. To further illus-
/ y [n] 2 (n - d x n) 4 results in the final v estimate. Un-
* n = nt
t O N-1 . (25)
trate this, let us consider the following like nt and vt obtained using multiple
linear combination:
/ y [n] 2 (n - d x n) 4 samples, A t given in (27) is based on
n=0
a single sample and, hence, can suffer
vt = tvt a + (1 - t) vt b, t ! (0, 1). (22) If we define A 2 /v p2 as the estimation from a large estimation error. Noticing
signal-to-noise ratio (SNR), where this, we suggest another refinement of
The mean square error (MSE) of the 2
v p is the power of the noise term t through minimizing the MSE, which
A
combined estimate can be computed as p [n] in (4), then the equality in (25) is calculated as
Table 1. A summary of the simulated methods, where the running time of each method is averaged over 105 independent trials.
n = 0, 1, f, N - 1
M3 New Estimate nt based on (26); estimate At using (27); perform the estimators in (18), attaining v a and v b ; 304.61
combine the two estimates in the linear manner depicted in (22), where the optimal t, as approximately
calculated in (25), is used as the combination coefficient; and refine At as done in (28).
M4 New and [5] Stage 1: Run M3 first, getting initial At , nt , and vt . 502.81
Stage 2: This is the same as in M2.
M5 [5] Perform the iterative WLS based on (9). 1,009.72
The simulations are run in MATLAB R2021a installed on a computing platform equipped with the Intel Xeon Gold 6238 R 2.2-GHz 38.5-MB L3 Cache (maximum turbo
frequency: 4 GHz; minimum: 3 GHz).
10–1
10–1 10–1
MSE
MSE
MSE
10–2
10–2 10–2
10–3
10–3 10–3
10–4
10 15 20 10 15 20 10 15 20
A2/σξ2 (dB) A2/σξ2 (dB) A2/σξ2 (dB)
(a) (b) (c)
M1 M2 M3 M4 M5
FIGURE 2. The MSEs of fitting results versus the SNR in the sampled Gaussian function with A = 1, where n and v are randomly generated based on
t , (b) nt , and (c) vt . Note that v 2p denotes the power of the noise term p [n] given in (4). The MSE is
the uniform distributions given in Table 2, for (a) A
5
calculated over 10 trials, each with independently generated and normally distributed noise.
10–1 10–1
10–1
MSE
MSE
MSE
10–2
10–2
10–2
2 4 6 8 10 12 2 4 6 8 10 12 2 4 6 8 10 12
Number of Iterations Number of Iterations Number of Iterations
(a) (b) (c)
M1 M2 M3 M4 M5
FIGURE 3. The MSEs of fitting results versus the number of iterations used in M5 and the second stage of M2/M4 for (a) At , (b) nt , and (c) vt , where
A = 1, A 2 /v p2 is 12 dB, and n and v are randomly generated based on the uniform distributions given in Table 2. The MSE is calculated over 105
trials, each with independently generated and normally distributed noise.
at the Epicenter of Ground-Shaking If you have any publication ideas I encourage authors to use Code Ocean
Research,” reports research results of for SPM, I encourage you to contact facilities (https://innovate.ieee.org/ieee
three projects focused on various issues the area editors or me (see the editorial -code-ocean/) for this purpose.
related to earthquakes, from forecasting board in the SPM web pages) to discuss
to localization of victims. your idea. Remember that the articles References
[1] R. Couillet, D. Trystram, and T. Menissier, “The
Among these articles, many are us- in SPM are not suited for the publish- submerged part of the AI-Ceberg [Perspectives],”
ing, at least partly, machine learning ing of either new results or surveys. IEEE Signal Process. Mag., vol. 39, no. 5, pp. 10–17,
Sep. 2022, doi: 10.1109/MSP.2022.3182938.
methods. Although some are consider- They are tutorial-like articles that must
[2] “IEEE Panel of Editors 2022.” IEEE.tv.
ing sparsity priors, I believe that use- be comprehensive for a wide audience Accessed: Apr. 29, 2022. [Online]. Available: https://
fulness, pros and cons, cost of these and accompanied by a relevant selec- ieeetv.ieee.org/channels/communities/welcome-from
-panel-of-editors-chair-poe-2022-0
methods (which are very greedy in pow- tion of both figures and references as
[3] “IEEE Publication Services and Products Board
er, computation, and memory), could be Robert Heath, the previous SPM editor- Operations Manual 2021,” IEEE Publications,
discussed more thoroughly. Benchmarks in-chief explained very clearly in [4]. Piscataway, NJ, USA, 2022. [Online]. Available: https://
pspb.ieee.org/images/files/files/opsmanual.pdf
with more classical methods should use Concerning the “Lecture Notes” and
[4] R. W. Heath, “Reflections on tutorials and sur-
metrics that are able to take into account Tips & Tricks” columns, I also think veys [From the Editor],” IEEE Signal Process. Mag.,
at least these parameters and not only a that sharing data and codes would be an vol. 37, no. 5, pp. 3–4, Sep. 2020, doi: 10.1109/
MSP.2020.3006648.
simple performance index. actual added value to these articles, and SP
across a number of academic publishing ing at Institut National Polytechnique de neering, speech processing, hyperspec-
companies. He joined IEEE initially as Grenoble, France. He has been a profes- tral imaging, and chemical engineering.
an editor for IEEE Press before transi- sor since 1989 and an emeritus profes- Since 2019, he has been a scientific
tioning to the Intellectual Property sor since 2019 at Université Grenoble advisor for scientific integrity for the
Rights (IPR) Office as an IPR specialist. Alpes, Grenoble 38402 France. Since 1979, French National Center of Scientific
Christian Jutten (christian.jutten@ his research has focused on statistical Research and, since January 2021, edi-
grenoble-inp.fr) received his master’s signal processing and machine learning tor-in-chief of IEEE Signal Processing
and Ph.D. degrees in electrical engineer- with applications in biomedical engi- Magazine. He is a Fellow of IEEE. SP
E
very individual perceives spatial anisms of human spatial hearing and techniques, specifically looking at
audio differently, due in large part communication. customization and personalization of
to the unique and complex shape of the audio rendering.
ears and head. Therefore, high-quality, Introduction ■■ It will explore, map, and model how
headphone-based spatial audio should Immersive audio is what we experi- the physical characteristics of spatial-
be uniquely tailored to each listener in ence in our everyday life, when we can ized auditory stimuli can influence
an effective and efficient manner. Arti- hear and interact with sounds com- observable behavioral, physiological,
ficial intelligence (AI) is a powerful ing from different positions around us. kinematic, and psychophysical reac-
tool that can be used to drive forward We can simulate this interactive audi- tions of listeners within social inter-
research in spatial audio personalization. tory experience within virtual reality action scenarios.
The SONICOM project aims to employ (VR) and augmented reality (AR) using ■■ It will evaluate the techniques devel-
a data-driven approach that links phys- off-the-shelf components such as head- oped and data-driven outputs in an
iological characteristics of the ear to phones, digital signal processors, iner- ecologically valid manner, exploit-
the individual acoustic filters, which tial sensors, and handheld controllers. ing AR/VR simulations as well as
allows us to localize sound sources and Immersive audio technologies have the real-life scenarios.
perceive them as being located around potential to revolutionize the way we ■■ It will create an ecosystem for audi-
us. A small amount of data acquired interact socially within AR/VR envi- tory data closely linked with model
from users could allow personalized ronments and applications. But several implementations and immersive
audio experiences, and AI could facili- major challenges still need to be tackled audio-rendering components, rein-
tate this by offering a new perspective before we can achieve sufficiently high- forcing the idea of reproducible
on the matter. A Bayesian approach quality simulations and control. This research and promoting future devel-
to computational neuroscience and will involve not only significant tech- opment and innovation in the area of
binaural sound reproduction will be nological advancements but also mea- auditory-based social interaction.
linked to create a metric for AI-based suring, understanding, and modeling
algorithms that will predict realistic low-level psychophysical (sensory) as Overview
spatial audio quality. Being able to con- well as high-level psychological (social SONICOM involves an international
sistently and repeatedly evaluate and interaction) perception. team of 10 research institutions and cre-
quantify the improvements brought by Funded by the Horizon 2020 FET- ative tech companies from six European
technological advancements, as well Proact scheme, the SONICOM project countries, all active in areas such as im-
as the impact these have on complex (www.sonicom.eu) started in January mersive acoustics, AI, spatial hearing,
interactions in virtual environments, 2021 and, over the course of the next auditory modeling, computational social
will be key for the development of five years, will aim to transform audi- intelligence, and interactive comput-
new techniques and for unlocking new tory social interaction and commu- ing. The workplan is centered around
approaches to understanding the mech- nication in AR/VR by achieving the three pivotal research work packages
following objectives: titled “Immersion,” “Interaction,” and
■■ It will design a new generation of “Beyond,” each introduced in one of the
Digital Object Identifier 10.1109/MSP.2022.3182929
Date of current version: 27 October 2022 immersive audio technologies and following sections. The first looks at
1053-5888/22©2022 Canadian Crown Copyright IEEE SIGNAL PROCESSING MAGAZINE | November 2022 | 85
immersive audio challenges dealing with and can be described by head-related based framework for the numerical cal-
the technical and sensory perspectives. transfer functions (HRTFs), which can culation of HRTFs. On the other hand,
On the other hand, the second focuses on be acoustically measured (e.g., see Fig- we focus on HRTF database match-
the interaction between these and higher- ure 1) or numerically modeled (e.g., in ing, an approach based on the hypoth-
level sociopsychological implications. [1]). Everyone perceives sound differ- esis that individuals can be paired with
Finally, the integration of core research, ently due to the particular shape of their existing high-quality HRTF data sets
proof-of-concepts evaluations, and cre- ears and head. For this reason, high- (measured or modeled) as long as they
ation of the auditory data ecosystem en- quality simulations should be uniquely share some relevant, predefined char-
sures that various outputs of the project tailored to each individual, effectively acteristics in the perceptual features
will have an impact beyond the end of and efficiently. Within SONICOM we space. To this end, we will expand the
SONICOM (see the “Beyond” section). propose a data-driven approach link- procedures based on objective similar-
ing the physiological characteristics of ity measures [4] and subjective listener
Immersion the ear to the individual acoustic filters input [5], [6]. The said measures con-
Before reaching the listener’s eardrums, employed for perceiving sound sources cern geometrical variations for PPMs,
the acoustic field is filtered due to shad- in space. perceptual deviations of the computed
owing and diffraction effects by the Our HRTF modeling research con- HRTFs, and signal-domain similarities
listener’s body, in particular, the head, siders a variety of approaches. On the for HRTF matching, all referenced to a
torso, and outer ears. This natural fil- one hand, we focus on the creation of project database comprising geometri-
tering depends on the spatial relation- parametric pinna models (PPMs) [2], cal scans and associated HRTF mea-
ship between the source and the listener [3] and their application to create an AI- surements from a set of individuals.
Being able to consistently and repeat-
ably evaluate and quantify the improve-
ments brought by these technological
advancements will be absolutely key,
not only for the development of new
techniques but also for unlocking new
approaches to understanding the mecha-
nisms of human spatial hearing. Our
approach to personalization presents a
new perspective on this problem, linking
Bayesian theories of active inference [7],
[8] and binaural (i.e., related to both ears)
sound reproduction to create data sets of
human behavior and perceptually valid
metrics modeling such behavior. By hav-
ing acoustic simulations validated against
acoustic measurements, and human
auditory models validated against actual
behavior, we will provide important tools
for the development of AI-based predic-
tors of realistic spatial audio quality.
We will also concentrate on the
issue of blending virtual objects in
real scenes, which is one of the corner-
stones of AR. To blend the real with
the virtual worlds in an AR scenario,
it is essential to develop techniques for
automatic estimation of the reverberant
characteristics of the real environment.
This will be achieved by character-
izing the acoustical environment sur-
rounding the AR user. The extracted
data can then be employed to generate
realistic virtual reverberation matching
the real world. After a set of pilot stud-
FIGURE 1. The HRTF measurement setup at Imperial College London, U.K. [24]. ies looking at perceptual needs in terms
Editor’s Note
Due to changing situations around
the world because of the corona-
virus (COVID-19) pandemic,
please double-check each confer-
ence’s website for the latest news
and updates.
©SHUTTERSTOCK.COM/G/SUNSINGER
2022
NOVEMBER
IEEE Workshop on Signal Processing
Systems (SiPS) The International Symposium of Biomedical Imaging will be held in Cartegena de Indias, Columbia,
2–4 November, Rennes, France. 18–21 April 2023.
General Chairs: John McAllister and Maxime Pelcat
URL: https://sips2022.insa-rennes.fr/
mathworks.com/machinelearning