0% found this document useful (0 votes)
16 views

IET Signal Processing - 2021 - Miron - Tensor Methods For Multisensor Signal Processing

This paper provides an overview of tensor-based methods for multisensor signal processing. It introduces tensor decompositions like the Tucker decomposition, canonical polyadic decomposition, and tensor-train decomposition. It describes estimation algorithms for these decompositions such as alternating least squares and high-order singular value decomposition. It then applies these tensor methods and algorithms to applications in direction-of-arrival estimation, multidimensional harmonic retrieval, and multiple-input–multiple-output wireless communication systems to illustrate their effectiveness.

Uploaded by

Mant wat
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views

IET Signal Processing - 2021 - Miron - Tensor Methods For Multisensor Signal Processing

This paper provides an overview of tensor-based methods for multisensor signal processing. It introduces tensor decompositions like the Tucker decomposition, canonical polyadic decomposition, and tensor-train decomposition. It describes estimation algorithms for these decompositions such as alternating least squares and high-order singular value decomposition. It then applies these tensor methods and algorithms to applications in direction-of-arrival estimation, multidimensional harmonic retrieval, and multiple-input–multiple-output wireless communication systems to illustrate their effectiveness.

Uploaded by

Mant wat
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

IET Signal Processing

Invited Paper

Tensor methods for multisensor signal ISSN 1751-9675


Received on 6th October 2020

processing
Revised 16th November 2020
Accepted on 24th November 2020
E-First on 2nd March 2021
doi: 10.1049/iet-spr.2020.0373
www.ietdl.org

Sebastian Miron1 , Yassine Zniyed1, Rémy Boyer2, André Lima Ferrer de Almeida3, Gérard Favier4,
David Brie1, Pierre Comon5
1Université de Lorraine, CNRS, CRAN, F-54000 Nancy, France
2Université de Lille, CNRS, CRIStAL, 59655 Lille, France
3Department of Teleinformatics Engineering, Federal University of Ceará, Fortaleza, CE 60440–900, Brazil
4Université de Côte d'Azur, CNRS, I3S Laboratory, Sophia Antipolis, France
5Université Grenoble Alpes, CNRS, GIPSA-Lab, 38000 Grenoble, France

E-mail: [email protected]

Abstract: Over the last two decades, tensor-based methods have received growing attention in the signal processing
community. In this work, the authors proposed a comprehensive overview of tensor-based models and methods for multisensor
signal processing. They presented for instance the Tucker decomposition, the canonical polyadic decomposition, the tensor-
train decomposition (TTD), the structured TTD, including nested Tucker train, as well as the associated optimisation strategies.
More precisely, they gave synthetic descriptions of state-of-the-art estimators as the alternating least square (ALS) algorithm,
the high-order singular value decomposition (HOSVD), and of more advanced algorithms as the rectified ALS, the TT-SVD/TT-
HSVD and the Joint dImensionally Reduction and Factor retrieval Estimator scheme. They illustrated the efficiency of the
introduced methodological and algorithmic concepts in the context of three important and timely signal processing-based
applications: the direction-of-arrival estimation based on sensor arrays, multidimensional harmonic retrieval and multiple-input–
multiple-output wireless communication systems.

1 Introduction 2 Tensor preliminaries: notations and operations


Tensors are, roughly speaking, generalisations of matrices to more Tensors are multilinear maps between two sets of linear spaces [9].
than two dimensions (orders larger than 2). Tensors present several Once these linear spaces are fixed as well as their bases, a tensor is
advantages over matrices, such as uniqueness of rank- characterised by its array of coordinates. In the present
decomposition under mild conditions, or preservation of the local contribution, we shall assume that this is the case and shall mainly
structure in N-way data processing. This is what motivated the work on array coordinates. Hence, with the usual abuse of
advent of tensors in data analysis (phonetics [1], psychometrics vocabulary, we shall assimilate tensors with their array
[2]) in the early 1970s and later on, in the signal processing representations. The number of dimensions (indices) of such an
community. Over the last 15 years approximately, the computer array is called the order of the tensor.
science community actively participated in the popularisation of Tensors are typeset with a bold calligraphic font (e.g., T),
tensors as data analysis tools. Due to their ability to model highly matrices in bold italic caps (e.g. Q); the elements of tensors or
structured data, tensors quickly became a staple of data mining and matrices are typeset as Ti, j, k and Qi, j, respectively; vectors are
machine learning fields (see, e.g. [3, 4]). denoted with bold italic lower cases (e.g. v), with entries vi. The
In multisensor signal processing, tensors are being used for
Frobenius norm of a matrix/tensor is denoted ∥ ⋅ ∥F; ( . )T, ( . )H
almost two decades in various applications such as direction of
arrival (DOA) estimation [5], harmonic retrieval [6], and ( . )† symbolise the matrix transpose, conjugate transpose and
communications systems [7] or radar [8]. However, despite a large the Moore–Penrose inverse, respectively, and ‘ ⋆ ’ denotes the
number of publications on this topic, we were unable to find in the convolution operator.
literature a comprehensive ‘point of entry’ article on the use of The tensor (outer) product between tensors, matrices, or
tensors in multisensor signal processing. The aim of this work is to vectors, is denoted by ⊗, as in e.g. T = v ⊗ M. The contraction on
fill this need. This paper does not intend to draw an exhaustive the pth index of a tensor is denoted as ∙ ; by convention, when
p
picture of the existing methods and models but rather provides an
contracted with a matrix, the summation always applies on the first
overview of the principles and main ideas used in tensor-based
index of the matrix if it appears after, or to the second index if it
multisensor signal processing. We illustrate our presentation by
appears first. In other words, the product A B between two
three popular multisensor applications: DOA estimation,
matrices can be written A ∙ B as well, without further index
multidimensional harmonic retrieval (MHR) and multiple-input–
multiple-output (MIMO) systems design. specification.
The paper is organised as follows: in Section 2 we define the For instance, the products Xi, j, k = ∑q Tq, j, k Mi, q,
main tensor concepts and notations and in Section 3 we introduce Yi, j, k = ∑q Ti, q, k M j, q and Zi, j, k = ∑q Ti, j, qMk, q can be written in
some of the most commonly used tensor decompositions and compact form as X = T ∙ M, Y = T ∙ M and Z = T ∙ M,
1 2 3
algorithms. These tensor decompositions are then applied to the
DOA estimation problem (Section 4), MHR (Section 5) and MIMO respectively. Note that we also have X = M T ∙ T.
1
wireless communication system (Section 6). Conclusions are The Kronecker product between two matrices is denoted
drawn in Section 7. A ⊠ B, and is defined by the block matrix:

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 693
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
A1, 1B A1, 2B … A1, J B Q
∏ Is
A2, 1B A2, 2B … A1, J B unfoldqT = reshape T′, Iq,
s=1
, (2)
A⊠B= , Iq
⋮ ⋮ ⋮ ⋮
AI , 1B AI , 2B … AI , J B
with
if the matrix A is of size I × J. Once the bases of linear spaces are
well defined, the tensor product between operators can be T′ = permute T, [q, 1, 2, …, q − 1, q + 1, q + 2, …, Q] ,
represented by Kronecker products [9]. The Khatri–Rao (column-
wise Kronecker) product between two matrices with the same where permute is a native MATLAB function that rearranges the
number J of columns is written as: dimensions of the Q-order tensor T.
A ⊙ B = a1 ⊠ b1, …a2 ⊠ b2, …aJ ⊠ bJ ,
3 Tensor decompositions and algorithms
if a j and b j represent the columns of A and B, respectively. We introduce hereafter the tensor decompositions (Section 3.1)
Analogously to the Frobenius norm of a matrix, the Frobenius used in the applications presented in Sections 4, 5 and 6, as well as
norm of a tensor T is defined as the square root of the sum of all some basic algorithms to compute these decompositions (Section
the squares of its elements, i.e.: 3.2).
The Tucker decomposition along with two of its variants (high-
∥ T ∥F = ∑ ∑ ∑ Ti j k . 2 order SVD and partial Tucker) are introduced in Section 3.1.1.
Tucker decomposition is a generic tensor tool that decomposes a
, ,
i j k
tensor into a set of non-structured factor matrices and a core tensor.
Contraction between two tensors A simple algorithm for computing the high-order SVD (Tucker
p decomposition with semi-unitary factor matrices) for third-order
The product ∙ between two tensors A and ℬ of size I1 × ⋯ × IQ tensors is given in Section 3.2.2. The canonical polyadic
q
decomposition (CPD) (Section 3.1.2) – probably the most widely
and J1 × ⋯ × JP, respectively, where Iq = J p, is a tensor of order
used tensor decomposition – can be seen as a Tucker
(Q + P − 2) denoted by: decomposition with a diagonal core. Its existence and uniqueness
p
issues are discussed in Section 3.1.3. The pseudo-code for the
[A ∙ ℬ]i1, , iq − 1, iq + 1, , iQ, j1, , jp − 1, jp + 1, , jP popular alternating least squares (ALSs) algorithm for estimating
q the CPD is given in Section 3.2.1. Section 3.1.4 introduces the
Iq tensor-train decomposition (TTD), a tool designed to efficiently
= ∑ [A]i [ℬ] j1, , jp − 1, k, jp + 1, , jP . handle high-order tensors (order higher than 3), by splitting them
1, , iq − 1, k , iq + 1, , iQ
k=1 into an interconnected set (‘train’) of lower-order tensors. A
particular class of TTD, the Structured Tensor-Train (STT) models,
Tensor reshaping is presented in Section 3.1.6. The tensor-train SVD (TT-SVD)
Tensor reshaping transforms a Q-order tensor T of dimensions algorithm for computing TTD is illustrated in Section 3.2.3. A link
I1 × ⋯ × IQ into a matrix T(q) having the product of the first q between the TTD and the CPD of high-order tensors (TT-CPD) is
dimensions of T, say I1⋯Iq, as the number of rows, and the developed in Section 3.2.4; the JIRAFE (Joint dImensionality
product of the remaining dimensions, Iq + 1⋯IQ, as the number of Reduction And Factors rEtrieval) method for estimating the
columns. In MATLAB, this reshaping can be obtained using the parameters of the TT-CPD model is also detailed. Two additional
tools used in MIMO application (Section 6), the least squares
native reshape function, such that
Kronecker (Section 3.2.5) and least squares Khatri–Rao (Section
q Q 3.2.6) factorisations, conclude this section.
T(q) = reshape T, ∏ Is, ∏ Is . (1)
s=1 s=q+1 3.1 Tensor decompositions

For example, let T be a fourth-order tensor of dimensions Any matrix can always be diagonalised by the congruent
2 × 2 × 2 × 2, defined by: transformation. In addition, the two linear transforms involved can
be imposed to be unitary: this is the singular value decomposition
1 3 5 7 (SVD). When we move from matrices to tensors, it is generally
T(: , : , 1, 1) = , T(: , : , 2, 1) = , impossible to use unitary transforms and obtain a diagonal tensor.
2 4 6 8 Consequently, depending on which property is desired, we end up
9 11 13 15 with two different decompositions. This is what is explained in the
T(: , : , 1, 2) = , T(: , : , 2, 2) = . next two subsections, where we limit ourselves to third-order
10 12 14 16
tensors to ease the presentation.
Matrix T(2) of dimensions 4 × 4 is then given by:
3.1.1 Tucker decomposition, HOSVD, and multilinear
1 5 9 13 rank: Given a tensor T of dimensions I × J × K, it is always
possible to find three matrices A, B and C of dimensions I × R1,
2 6 10 14
T(2) = reshape T, 4, 4 = . J × R2 and K × R3, respectively, such that
3 7 11 15
4 8 12 16 R1 R2 R3
Ti, j, k = ∑ ∑ ∑ Ai mB j nCk p Gm n p,
, , , , , (3)
The so-called flattening or unfolding operation is a special case of m=1n=1 p=1
the reshaping operation [9]. The most used flattenings are those
keeping one of the original dimensions, say Nq, as the number of where R1, R2, R3 are minimal and R1 ≤ I, R2 ≤ J, R3 ≤ K. Tensor G,
rows. Hence, the so-called qth matrix unfolding of T is of often called core tensor, is of dimensions R1 × R2 × R3. The triplet
dimension Iq × I1…Iq − 1 Iq + 1 …IQ. Compared to the flattening of minimal values of R1, R2, R3 forms the multilinear rank of T.
operation, the reshape function can be considered as a more This is referred to as the Tucker decomposition of T, and can be
general flattening operation, in the sense that written in a more compact way as:

T= A, B, C; G . (4)
694 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
Each Rℓ is uniquely defined, and corresponds to the rank of the The uniqueness of the CPD (8) should not be confused with its
linear operator associated with the ℓth matrix unfolding of T, array representation (7). In fact, the latter is never unique, since
unfoldℓT. The point is that the Tucker decomposition is not expressing a rank-1 tensor as the tensor product of N vectors is
unique, for it is defined up to three invertible matrices subject to N − 1 scaling indeterminacies. For instance,
M1, M2, M3 , because: D = a ⊗ b ⊗ c can also be written as αa ⊗ βb ⊗ γc provided
αβγ = 1. This is precisely the difference between a tensor space
A, B, C; G = AM1−1, BM2−1, CM3−1; G′ and a product of linear spaces [9, 13, 14]. This is the reason why
the wording of ‘essential uniqueness’ is sometimes found in the
literature; it expresses the fact that there are N − 1 scaling
if G′ = M1, M2, M3; G . Yet, as in the SVD of matrices, it is
indeterminacies, and that the order of summation is subject to
possible to impose that the core tensor is obtained via semi-unitary
permutation because the addition is commutative.
transforms {U, V, W} as:
With Definition (2), the CPD can be rewritten in matrix form.
For a third-order rank-R tensor T = A, B, C; S for example, the
T = U, V, W; G , (5)
three flattened representations of the CPD are [15]:
where UHU = IR1, V HV = IR2 and W HW = IR3. Equation (5)
unfold1T = AS(C ⊙ B)T, (9)
defines the high-order SVD (HOSVD), sometimes referred to as
multilinear SVD of tensor T [10].
unfold2T = BS(C ⊙ A)T, (10)
Partial Tucker decomposition: We present next a variant of the
Tucker decomposition for an Q-order tensor X ∈ ℂI1 × ⋯ × IQ, with
unfold3T = CS(B ⊙ A)T, (11)
factor matrices A(q) ∈ ℂIq × Rq whose Q − Q1 last ones are equal to
identity matrices IIq of order Iq, for q = Q1 + 1, ⋯, Q. This so- where S denotes the diagonal matrix with entries σ(r) = Sr, r, r,
called Tucker-(Q1, Q) decomposition is compactly written as ∀1 ≤ r ≤ R.
A(1), …, A(Q1), IIQ1 + 1, …, IIQ; G , where the core tensor G is of One of the major interests in the CPD (8) lies in its uniqueness.
In particular, it allows to relax the orthogonality constraint
dimensions R1 × ⋯ × RQ1 × IQ1 + 1 × ⋯ × IQ, which induces Rq = Iq
(necessary in the case of matrices to achieve some form of
for q = Q1 + 1, …, Q. See [11] for more details. For instance, the uniqueness), which is often incompatible with physical reality.
standard Tucker2 and Tucker1 decompositions correspond to More precisely, several uniqueness conditions have been derived in
(Q1, Q) = (1, 3) and (Q1, Q) = (2, 3), respectively. the literature. We shall quote two of them. A sufficient condition is
that [16–18]:
3.1.2 Exact CP decomposition, rank, uniqueness: Of course,
there is no reason that tensor G in (5) be diagonal; this is clear by 2R ≤ krank{A} + krank{B} + krank{C} − 2, (12)
just looking at the number of degrees of freedom of both sides. On
the other hand, if the unitary constraint of matrices {U, V, W} is where krank{A} denotes the so-called Kruskal rank of A [9, 16];
relaxed, it is possible to decompose any tensor T as: by definition [A matrix M has Kruskal rank R if any subset of R
columns forms a full rank matrix. Remember that M has rank R if
T= A, B, C; S , (6) there exists at least one subset of R columns forming a full rank
matrix.], the krank is always smaller than or equal to the usual
where tensor S is diagonal. This is known as the Canonical matrix rank. If entries of A are seen as randomly drawn according
Polyadic (CP [The acronym CP also stands for Candecomp/Parafac to an absolutely continuous distribution, krank{A} can be replaced
in the literature, due to two contributions [1, 2] in which this tensor by the smallest of its two dimensions. In the latter case, the
decomposition has been independently rediscovered, years after the condition is referred to as generic [19, 20].
original publication [12].]) decomposition (CPD). Note that now, A second generic condition ensuring uniqueness of the CPD is
not only matrices A, B, C are not semi-unitary, but their number given by the bound of the so-called expected rank [21–23]:
of columns, R, may exceed the number of their rows. The minimal
value of R such that (6) holds exactly is called the rank of the R1R2R3
R≤ − 1, (13)
tensor T. The explicit writing of (6) in terms of array entries is: R1 + R2 + R3 − 2

R where Rℓ are the multilinear ranks defined in Section 3.1.1. To sum


Ti, j, k = ∑ Ai rB j rCk r σ(r),
, , , (7) up, the exact CP decomposition (8) is almost surely unique if the
r=1
tensor rank is not too large.
if σ(r) ’s denote the diagonal entries of S. But clearer writing of
this CPD is provided in terms of decomposable tensors [9] as 3.1.3 Approximate CPD, existence, uniqueness: Things
follows: become more complicated when it is thought to approximate a
tensor by another one of lower rank. To fix the ideas, suppose that
R our objective is to minimise:
T= ∑ σrD(r), (8)
Ro 2
r=1
Υ=∥T− ∑ σ(r) a(r) ⊗ b(r) ⊗ c(r) ∥ ,
where D(r) = a(r) ⊗ b(r) ⊗ c(r), that is, D(r)i, j, k = Ai, r B j, rCk, r, r=1 F
∥ D(r) ∥ = 1, and σr > 0. In fact, decomposable tensors may be
seen as rank-1 tensors. This writing is clearer because it does not where σ(r) > 0 and a(r) = b(r) = c(r) = 1,
depend on the way tensors D(r) are represented – and their ∀r, 1 ≤ r ≤ Ro. Unfortunately, it has been pointed out in e.g. [24]
representation is precisely rarely unique, which is not related to the that this objective function may not always have a minimum (but
uniqueness of (8). Note that the definition of tensor rank is pretty only an infimum), which means that this low-rank approximation
much the same as matrix rank. Hence, this CPD may be seen as problem is ill-posed; there are ways to circumvent the difficulty,
another natural generalisation of the matrix SVD to tensors, the for instance by imposing extraneous angular constraints [25],
former being HOSVD seen in the previous section. Because the which is meaningful in the context of DOA estimation for instance.
acronym CP is sometimes used to mean ‘Completely Positive’, it On the other hand, if all objects are real non-negative, then the
may sometimes be more suitable to call it rank decomposition, problem becomes well-posed [26]. In the latter case, it is important
when the number of terms R is indeed minimal. to recall that imposing a non-negative constraint generally

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 695
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
and the TT-ranks are all equal to the canonical rank R.

Proof: The TTD of the Q-order identity tensor ℐQ, R is given by

1 1 1 1
ℐQ, R = IR ∙ ℐ3, R ∙ ⋯ ∙ ℐ3, R ∙ IR . (22)
2 3 Q−1 Q
Fig. 1 TT decomposition of a Q-order tensor
Substituting the above TTD into the CPD, we get
increases the rank [9], even for matrices. This being said, it has
been proved in [22] that the best non-negative low-rank tensor 1 1 1

approximation of a non-negative tensor is almost surely unique. X = (IR ∙ ℐ3, R ∙ ⋯ ∙ IR) ∙ P1… ∙ PQ (23)
2 3 Q 1 Q
One can even look at the conditions under which the best low-rank
approximation admits a unique CPD [27], but this is more by expressing the entries of X and reorganising them, we can
involved. equivalently write
To conclude, the best low-rank approximate of a tensor does not
always exist, and this is still often ignored in the literature. 1 1 1
X = (P1 IR) ∙ (ℐ3, R ∙ P2) ∙ ⋯ ∙ (IR PQT ) . (24)
2 2 3 Q
3.1.4 Tensor-train decomposition and TT-ranks: A Q-order
tensor of size I1 × ⋯ × IQ admits a decomposition into a train of By identifying the TT-cores in (24) with those in (14), we can
tensors [28] if deduce the relations (19), (20) and (21). Then, it is straightforward
to conclude that the TT-ranks are all identical and equal to the
1 1 1 1
X = G(1) ∙ G(2) ∙ G(3) ∙ … ∙ G(Q − 1) ∙ G(Q),
1
(14) canonical rank R. □
2 3 4 Q−1 Q Thus, conditionally to the knowledge of the TT-cores, it is
theoretically possible to recover the factors of the CPD by a one-to-
where the TT-cores G(1), G(q)(2 ≤ q ≤ Q − 1), and G(Q) are, one methodology. In Section 3.2.4, we present the so-called
respectively, of size I1 × R1, Rq − 1 × Iq × Rq and RQ − 1 × IQ. The JIRAFE framework, used for that aim.
Q − 1 dimensions {R1, , RQ − 1} are referred to as the TT-ranks with
boundary conditions R0 = RQ = 1. 3.1.6 Structured TT models: In this subsection, we present a
The idea behind the TTD is to transform a high Q-order tensor particular class of TT models, the so-called STT models. These
into a set of much lower third-order tensors, which allows to break models are composed of a train of third-order tensors, each tensor
the ‘curse of dimensionality’ [29]. Indeed, just like the CPD, the being represented by means of a tensor decomposition like CP,
storage cost of the TTD scales linearly with the order Q. A graph- Tucker or generalised Tucker.
based representation of the TTD of a Q-order tensor X is given in For a Q-order tensor X ∈ ℂI1 × I2⋯ × IQ, an STT model can be
Fig. 1. It is straightforward to see that the TTD is not unique since written as:
[30]
1 1 1
T
X = T(1) ∙ T(2) ∙ ⋯ ∙ T(Q − 2), (25)
[X]i1, ⋯, iQ = a1(i1) A2(i2)⋯AQ − 1(iQ − 1)aQ(iQ) (15) 3 4 Q−1

where with T(1) ∈ ℂI1 × I2 × R2, T(Q − 2) ∈ ℂR2Q − 6 × IQ − 1 × IQ, and


(q)
T ∈ℂ R2q − 2 × I q+1 × R2q
for 2 ≤ q ≤ Q − 3.
a1(i1) = M1[G(1)]iT1, : : R1 × 1, (16) For instance, in the case of a fifth-order tensor X ∈ ℂI1 × I2⋯ × I5,
let us assume that T(q) satisfies a Tucker-(1,3) decomposition for
aQ(iQ) = MQ−1− 1[G(Q)] : , iQ : RQ − 1 × 1, (17) q ∈ {1, 2} while T(3) satisfies a Tucker-(2,3) decomposition such
that [11]:
Aq(iq) = Mq− T− 1[G(q)] : , iq, : Mq : Rq − 1 × Rq, (18)
T(1) = A(1), II2, IR2; G(1) ∈ ℂI1 × I2 × R2 (26)
with Mq, an invertible matrix of size Rq × Rq. In [30], it is shown
that, when we apply decomposition algorithms such as TT-SVD or T(2) = A(2), II3, IR4; G(2) ∈ ℂR2 × I3 × R4 (27)
TT-HSVD, the matrices Mq are change-of-basis matrices, that are
mainly linked to the estimation of dominant subspaces by the SVD. T(3) = A(3), II4, A(4); G(3) ∈ ℂR4 × I4 × I5 (28)
3.1.5 Model equivalence between TTD and CPD: An
with the core tensors
interesting model equivalence property between a Q-order CPD of
rank-R and a train of (Q − 2) third-order CPD(s) of rank R and two G(1) ∈ ℂR1 × I2 × R2, G(2) ∈ ℂR3 × I3 × R4, G(3) ∈ ℂR5 × I4 × R6, and the matrix
matrices is briefly discussed in [28]. We can summarise this factors A(1) ∈ ℂI1 × R1, A(2) ∈ ℂR2 × R3, A(3) ∈ ℂR4 × R5, A(4) ∈ ℂI5 × R6. We
fundamental algebraic relation according to the following result. then obtain a structured Tucker train of order five, abbreviated as
STuT(5), which can be written by means of the following scalar
Theorem 1: If the tensor X follows a Q-order CPD of rank-R equation:
with factors Pq, then a TTD can be given by:
R1 R2 R6

G1 = P1, (19) xi1, i2, i3, i4, i5 = ∑ ∑ ⋯ ∑ ai (1) (1)


g (2)
a g(2)
1, r1 r1, i2, r2 r2, r3 r3, i3, r4
r1 = 1 r2 = 1 r6 = 1 (29)
(see (20)) × ar(3)4, r5gr(3)5, i4, r6ai(4)
5, r6
,

GQ = PQT , (21) or compactly as:

Gq = ℐ3, R ∙ Pq (third − order CPD), where 2 ≤ q ≤ Q − 1, (20)


2

696 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
R3 R4
tr(2)2, i3, i4 = ∑ ∑ gr (41)
(2) (2) (3)
a a
3, i3, r4 r2, r3 i4, r4
r3 = 1 r4 = 1

R1 R2
ti(3)
1, i2, r3
= ∑ ∑ gr (1) (1)
a (2)
a
1, i2, r2 i1, r1 r2, r3
(42)
r1 = 1 r2 = 1
Fig. 2 NTT(4) model for a fourth-order tensor
R4
tr(4)3, i3, i4 = ∑ gr (43)
(2) (3)
a
3, i3, r4 i4, r4
.
r4 = 1

The NTT(4) model (31) can be viewed as the nesting of the third-
order Tucker models T(3) and T(2) which share the matrix factor
A(2). It can also be interpreted as a contraction of a Tucker-(1,3)
model with a Tucker-(2,3) model, along their common mode:
1 1
Fig. 3 Nested CP model for a fourth-order tensor X = T(1) ∙ T(2) = T(3) ∙ T(4) . (44)
3 3
1 1
X = T(1) ∙ T(2) ∙ T(3) . (30) These contraction operations correspond to summing the entries of
3 4
the third-order tensors T(1) and T(2), or T(3) and T(4), along their
Comparing (29) and (30) of the STuT(5) model with (14) of the TT common mode r2, or r3, as follows:
model with Q = 5, we can conclude that the STuT model R2
corresponds to a TT model for which each tensor of the train
∑ ti (45)
(1) (2)
satisfies a Tucker-(1,3) or Tucker-(2,3) decomposition. xi1, i2, i3, i4 = t
1, i2, r2 r2, i3, i4
r2 = 1
Such an STuT model was derived for the fourth-order tensor of
received signals in a cooperative wireless MIMO communication R3
system [31], and then generalised to a Q-order tensor in the case of (46)
multi-hop MIMO relay systems [32]. This model can also be
= ∑ ti (3)
t(4)
1, i2, r3 r3, i3, i4
.
r3 = 1
interpreted as a nested Tucker train (NTT) model, with two
successive tensors sharing a common factor.
Defining the fourth-order tensor C ∈ ℂR1 × I2 × I3 × R4 such that:
Thus, in the case of a fourth-order tensor X ∈ ℂI1 × I2 × I3 × I4, as
illustrated by means of Fig. 2, with R2 R3
A(1) ∈ ℂI1 × R1, G(1) ∈ ℂR1 × I2 × R2, A(2) ∈ ℂR2 × R3, G(2) ∈ ℂR3 × I3 × R4, A(3), cr1, i2, i3, r4 = ∑ ∑ gr (1)
a(2) (2)
g
1, i2, r2 r2, r3 r3, i3, r4
, (47)
∈ ℂI4 × R4 r2 = 1 r3 = 1
the NTT(4) model can be described by means of the following
scalar equation: the NTT(4) model (31) can also be interpreted as the following
Tucker-(2,4) model [33]:
R1 R2 R3 R4
xi1, i2, i3, i4 = ∑ ∑ ∑ ∑ ai (1)
g(1) (2)
a g(2)
a (3)
1, r1 r1, i2, r2 r2, r3 r3, i3, r4 i4, r4
. (31) X = C ∙ A(1) ∙ A(3), (48)
r1 = 1 r2 = 1 r3 = 1 r4 = 1 1 4

Let us define the following third-order Tucker models: or in the scalar form:
R1 R4
T(1) = A(1), II2, IR2; G(1) (32)
xi1, i2, i3, i4 = ∑ ∑ cr 1, i2, i3, r4
ai(1) a(3) .
1, r1 i4, r4
(49)
r1 = 1 r4 = 1
(1) (1) I1 × I2 × R2
=G ∙A ∈ℂ (33)
1
The nested CP model introduced in [34], and exploited in [35] in
the context of MIMO relay systems, is a special case of the nested
(34)
(2) (2) (3) (2)
T = A , II3, A ; G Tucker model (31), defined as follows:

= G(2) ∙ A(2) ∙ A(3) ∈ ℂR2 × I3 × I4 (35)


R1 R2
∑ ∑ ai (50)
(1) (1) (2) (2) (3)
1 3 xi1, i2, i3, i4 = g a g a
1, r1 i2, r1 r1, r2 i3, r2 i4, r2
,
r1 = 1 r2 = 1
T(3) = A(1), II2, A(2); G(1) (36)
with A(1) ∈ ℂI1 × R1, G(1) ∈ ℂI2 × R1, A(2) ∈ ℂR1 × R2, G(2) ∈ ℂI3 × R2, A(3).
= G(1) ∙ A(1) ∙ A (2)T
∈ ℂI1 × I2 × R3 (37) ∈ ℂI4 × R2
1 3 The nested CP model is illustrated in Fig. 3 for a fourth-order
tensor.
T(4) = IR3, II3, A(3); G(2) (38) This nested CP model can be deduced from the nested Tucker
model (31) using the following correspondences:
= G(2) ∙ A(3) ∈ ℂR3 × I3 × I4 . (39)
3 (r1, r2, r3, r4) ↔ (r1, r1, r2, r2) (51)

These Tucker models can be written in an element-wise form as: (A(1), G(1), A(2), G(2), A(3)) ↔ (A(1), G(1), A(2), G(2), A(3)) . (52)
R1
More generally, one can define a STT model (25) for which each
ti(1) ∑ gr (40)
(1) (1)
1, i2, r2
= a
1, i2, r2 i1, r1 elementary tensor of the train has a CP decomposition. Thus, for a
r1 = 1
Q-order tensor X, (25) then becomes a structured CP train (SCPT)
such that:

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 697
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
T
T(1) = A(1), G(1), A(2) ; ℐR1 ∈ ℂI1 × I2 × R2
T
T(q) = A(q), G(q), A(q + 1) ; ℐRq

∈ ℂRq − 1 × Iq + 1 × Rq + 1 for 2 ≤ q ≤ Q − 3
T(Q − 2) = A(Q − 2), G(Q − 2), A(Q − 1); ℐRQ − 2
∈ ℂRQ − 2 × IQ − 1 × IQ,

with the matrix factors A(1) ∈ ℂI1 × R1, A(q) ∈ ℂRq − 1 × Rq, for
2 ≤ q ≤ Q − 1, A(Q − 1) ∈ ℂIQ × RQ − 2, and G(q) ∈ ℂIq + 1 × Rq, for Fig. 4 Algorithm 1 ALS algorithm
1 ≤ q ≤ Q − 2. The SCPT model of order Q can also be written in
the following scalar form:
R1 RQ − 2
xi1, ⋯, iQ = ∑⋯ ∑ ai(1) g(1) a(2) g(2)
1, r1 i2, r1 r1, r2 i3, r2
r1 = 1 rQ − 2 = 1

× ar(3)2, r3gi(3)
4, r3
⋯gi(QQ−−12), rQ − 2ai(QQ, r−Q1)− 2 .

3.2 Algorithms
3.2.1 ALSs algorithm: There are many algorithms to compute a
CP decomposition. By far, the most used one is the ALSs. ALS
was proposed in [1, 2], and is considered today as the ‘workhorse’
for CPD computation. It is a simple algorithm that fixes, iteratively,
all but one factor, which is then updated by solving a linear least
squares problem. In Algorithm 1 (Fig. 4), we present the ALS
algorithm for a third-order tensor T ≈ A, B, C; ℐ3, R . The
generalisation for a Q-order case is straightforward.
Generally CritStop is based on the evaluation of the fitting
error ∥ T − A, B, C; ℐ3, R ∥F.

3.2.2 High-order SVD algorithm: Based on the definition of the


HOSVD given in (5), an algorithm to compute the HOSVD is
presented in Algorithm 2. This algorithm was proposed in [10], and
it relies on the computation of the left singular vectors of the
unfoldings unfoldqT. The truncated HOSVD has a moderate
computing cost, and is not necessarily optimal, but it gives a good
approximation, which can be used as initialisation for ALS-like
algorithms. For ease of presentation, we derive the method only for
a third-order tensor, but the generalisation to a Q-order tensor is
straightforward. Fig. 5 TT-SVD algorithm applied on a fourth-order tensor

Algorithm 2 HOSVD algorithm: It is worth noting that the TT-SVD in its current state cannot be
parallelised, which is a problem when we deal with very high order
Input: Third-order tensor T, multilinear rank {R1, R2, R3} tensors. An alternative method to compute the TTD is to consider
Output: U, V, W and G different reshapings than those considered in the TT-SVD. Indeed,
(1) U ← R1 first left singular vectors of unfold1T the recently proposed TT-HSVD [30] algorithm, for tensor-train
(2) V ← R2 first left singular vectors of unfold2T hierarchical SVD, is a hierarchical algorithm for TTD, that
suggests to combine more than one mode in each dimension of the
(3) W ← R3 first left singular vectors of unfold3T reshaped matrices, i.e. instead of considering the matrix X(1) of size
(4) G = T ∙ U H ∙ V H ∙ W H I1 × I2I3I4 , we can for example use as first unfolding the matrix
1 2 3
X(2) of size I1I2 × I3I4 . Note that using this strategy allows to
parallelise the decomposition, after each SVD, across several
3.2.3 TTD computation with the TT-SVD algorithm: The TT- processors, which suits very well high order tensors. Moreover, the
SVD algorithm has been introduced in [28]. This algorithm choice of the best reshaping strategy when the order is very high is
minimises in a sequential way the following LS criterion: discussed in [30] in terms of the algorithmic complexity. Indeed,
2
[30] shows that reshapings with the ‘most square’ matrix (i.e.
1 1 1 1
ψ(X) = ∥ X − A1 ∙ A2 ∙ ⋯ ∙ AQ − 1 ∙ AQ ∥ . (53) matrices with more balanced row and column dimensions), leads to
2 3 Q−1 Q F the lower computational complexity.
Remark that from an estimation point of view, the TT-SVD and
In Fig. 5, we present in a schematic way the TT-SVD algorithm the TT-HSVD algorithms suffer from the lack of uniqueness of the
applied on a fourth-order tensor X. As illustrated, the TT-SVD TT-cores described in Section 3.1.4. Indeed, as the latent matrices
computes in a sequential way Q − 1 SVDs on matrix reshapings of M1, …, MQ − 1 are unknown, the true TT-cores remain unknown.
the Q-order original tensor. Note that at each step, we apply the In the next section, we propose a methodology to solve this
truncated SVD on the reshaped matrices V(2) (q)
of size problem in the important context where the observed tensor
follows a Q-order CPD of rank R.
RqIq + 1 × Iq + 2⋯IQ to recover matrices U (q + 1)
and V (q + 1), this
latter containing the product of the pseudo-diagonal singular values
3.2.4 JIRAFE principle for CPD: In this section, we present a
matrix and the right singular vectors matrix.
TT-based methodology [36], JIRAFE, for high-order CPD factors

698 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
min ∥ A1 − P1 M1−1 ∥F + ∥ AQ − MQ − 1 PQ ∥F (57)
M, ℙ

Q−2
+ ∑ ϕ(Aq) , (58)
q=2

where M = {M1, …, MQ} and ℙ = {P1, …, PQ} and

ϕ(Aq) = ∥ Aq − ℐ3, R ∙ Mq − 1 ∙ Pq ∙ Mq− T ∥2F . (59)


Fig. 6 Third-order CPD of the qth TT-core 1 2 3

In Algorithm 3, we denote by ϕ(A M) the criterion ϕ(A) when


factor M has been previously estimated. The JIRAFE approach
breaks this delicate optimisation problem into a set of Q low-
dimensional and coupled LS criterion. Indeed, the JIRAFE
methodology is essentially based on a single tridimensional LS
optimisation, Q − 3 bidimensional LS optimisations and the
computation of two inverse R × R matrices. It is worth noting that
any processing applicable to CPDs can be done at the level of the
estimated third-order CPD TT-cores, using the JIRAFE principle,
with a much lower complexity. Moreover, the ambiguities of the
CPD-train model, presented in Theorem 2, are the same as the ones
of the CPD. Indeed, in [36], a proof of the following result is
given:

Remark 2: Just like the CPD, the CPD-Train model, given in


Theorem 2, is characterised by the following indeterminacies:
Fig. 7 Algorithm 3 JIRAFE for CPD (i) A unique column permutation matrix denoted by Π.
(ii) Diagonal scaling matrices satisfying the following relation:
retrieval. The acronym JIRAFE stands for ‘Joint dImensionality
Reduction And Factors rEtrieval’ and the method is described in Λ1Λ2Λ3⋯ΛQ − 1ΛQ = IR (60)
[36, 37] by the following two-step procedure:

(i) Reduce the dimensionality of the original factor retrieval where Λk is the scaling ambiguity for the k-mode factor Pk.
problem by breaking the difficult multidimensional optimisation
problem into a collection of simpler optimisation problems on 3.2.5 Least squares Kronecker factorisation: Consider the
small-order tensors. This step is performed using the TT-SVD following minimisation problem
algorithm.
(ii) Design a factor retrieval strategy by exploiting (or not) the min∥ X − A ⊠ B ∥F, (61)
A, B
coupled existing structure between the first and third factors of two
consecutive TT-cores. Here, the goal is to minimise a sum of
coupled least-square (LS) criteria. where A ∈ ℂI2 × R2, B ∈ ℂI1 × R1 and X = A ⊗ B + V ∈ ℂI1I2 × R1R2,
and V represents a zero-mean uncorrelated noise term. The
In [36], the structures of the TT-cores associated with the TT-SVD solution of the problem in (62), is based on a rank-one matrix
algorithm are described and a family of estimation algorithms is approximation (via SVD) of X (a permuted version of X, the
proposed. This methodology is based on the following result. construction of which was proposed in [38]). The problem in (62)
becomes
Theorem 2: If the data tensor follows a Q-order CPD of rank R
parametrised by Q full column rank factors P1, …, PQ . The TT- min∥ X¯ − b ⊗ a ∥F, (62)
a, b
SVD algorithm recovers the TT-cores such that:
meaning to find the nearest rank-one matrix to X¯ , where
A1 = P1 M1−1, (54)
a = vec(A) ∈ ℂI2R2 × 1 and b = vec(B) ∈ ℂI1R1 × 1. In [39], the
authors proposed a solution generalising [38] to a Kronecker
Aq = ℐ3, R ∙ Mq − 1 ∙ Pq ∙ Mq− T, where 2 ≤ q ≤ Q − 1 (55) product involving N factor matrices. Let us consider the case
1 2 3
N = 3, usually encountered in practice. The problem then becomes
AQ = MQ − 1 PQT , (56)
min∥ X − A ⊠ B ⊠ C ∥F, (63)
A, B
where Mq is a non-singular R × R change of basis matrix.
This means that if a Q-order tensor admits a rank-R CPD, then where A ∈ ℂI3 × R3, B ∈ ℂI2 × R2 and C ∈ ℂI1 × R1. The problem in
its TTD involves a train of (Q − 2) third-order CPD(s) and has all (64) now becomes
identical TT-ranks such as R1 = ⋯ = RQ − 1 = R. The factors can be
derived straightforwardly from the TT-cores up to two change-of- min ∥ X̄ − c ⊗ b ⊗ a ∥F, (64)
basis matrices. a, b, c

Remark 1: Note that each TT-core for 2 ≤ q ≤ Q − 1 follows a where a = vec(A) ∈ ℂI3R3 × 1, b = vec(B) ∈ ℂI2R2 × 1,
CPD coupled with its adjacent TT-cores (see Fig. 6). c = vec(C) ∈ ℂI1R1 × 1. We have that x̄ = vec X¯ and
The above theorem and the remark allow us to derive the factor
retrieval scheme, called JIRAFE and presented in Algorithm 3 X = T x̄ ∈ ℂI1R1 × I2R2 × I3R3, where the operator T{ ⋅ } maps the
(Fig. 7) as a pseudo-code which minimises the following criterion: elements of x̄ into X̄, as follows:

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 699
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
x̄q1 + (q2 − 1)Q1 + (q3 − 1)Q1Q2 → X̄q1, q2, q3 (65) approximation problem R times in parallel, one for each column xr,
T{ ⋅ } r = 1, …, R.
where qi = {1, …, Qi} and Qi = IiRi, with i = {1, 2, 3}. Otherwise
4 Tensor-based DOA estimation with sensor
stated, we have X = reshape x, Q1, Q2, Q3 . Hence, finding the
matrix triplet A, B, C that solves (63) is equivalent to finding the arrays
vector triplet a, b, c that solves (64), i.e. the solution of a A fundamental problem in array processing is the estimation of the
Kronecker approximation problem can be recast as the solution to a DOAs for multiple sources impinging on an array of sensors.
rank-one tensor approximation problem, for which effective Assume that P narrow-band far-field sources are impinging on an
algorithms exist in the literature (see, e.g. [40–42]). array of L identical sensors (P < L, in general). The direction of
arrival of a source p in a Cartesian coordinate system OXYZ
3.2.6 Least squares Khatri–Rao factorisation: Consider the associated with the array is given by the unitary vector
following minimisation problem T
kp = sin θ pcos ϕp sin θ psin ϕp cos θ p , where θ and ϕ are the
elevation and azimuth angles, respectively, as illustrated on Fig. 8.
min∥ X − A ⊙ B ∥F, (66) With these notations, a snapshot of the array at time instant t is
A, B
given by:
where A ∈ ℂI × R, B ∈ ℂJ × R and X = A ⊙ B + V ∈ ℂJI × R, and V P
represents the zero-mean uncorrelated noise term. Note that y(t) = ∑ a(kp)sp(t) + b(t) = Ax(t) + b(t), (70)
problem (66) can be rewritten as p=1

R where y(t) is the (L × 1) output vector of the array,


min ∑ ∥ xr − ar ⊠ br ∥2, (67) A = a(k1), …, a(kP) is the (L × P) steering vectors matrix of the P
ar, br i = 1
sources, x(t) = [s1(t), …, sP(t)]T is the (P × 1) source signals vector
or, equivalently, at time instant t, and b(t) is a (L × 1) vector that accounts for the
spatially and temporally white Gaussian noise on the array sensors.
R If K snapshots t1, t2, …, tK are considered, the output of the array
min ∑ ∥ X¯ r − br ⊗ ar ∥F, (68) can be expressed in the matrix form as:
ar, br i = 1

Y = [y(t1), …, y(tK )]
where X¯ r ∈ ℂJ × I . Defining UrΣr BrH as the singular value
= A(k1, …, kP)[x(t1), …, x(tK )] + B (71)
decomposition (SVD) of Xr, estimates for ar ∈ ℂI and br ∈ ℂJ , T
(r = 1, …, R) can be obtained by truncating the SVD of Xr to its = A(k1, …, kP)S + B
dominant eigenmodes, i.e. [43]
with S = s1, …, sP , a K × P matrix gathering on its columns the
T
^ (1) (1)
br = σr ur and ar = σr vr , ^ (1) (1) (1) ∗
(69) K time samples for the P sources, sp = sp(t1), …, sp tK , with
p = 1, …, P and B = b(t1), …, b tK , a (L × K) noise matrix. The
where ur(1) ∈ ℂJ and vr(1) ∈ ℂI are the dominant first left and right DOA estimation problem consists in finding k1, …, kP from the
singular vectors of Ui and Vi, respectively, while σr(1) denotes the data Y.
The use of multilinear algebra to solve this problem is inspired
largest singular value of X¯ r. Hence, estimates of the full matrices A
by the principle of ESPRIT algorithm introduced by Roy et al. in
and B that minimise (67) are obtained by repeating such a rank-1 [44, 45]. The idea is to exploit the invariances in the array output
data in order to create multilinear structures. These invariances can
be intrinsic to the physics of the acquired signals (i.e. polarisation),
to the acquisition setup (i.e. spatially shifted subarrays), or
artificially created (i.e. matricisation/time–frequency transform of
1D signals).
The main idea of ESPRIT, when applied to DOA estimation is
to employ two identical subarrays, the second one being spatially
shifted compared to the first one by a known displacement vector
δ, as illustrated in Fig. 9. The outputs of the first and the second
subarrays, denoted by y1(t) and y2(t), respectively, can then be
expressed as:

Fig. 8 DOA estimation set-up y1(t) = Ax(t) + b1(t), (72)

y2(t) = AΦx(t) + b2(t), (73)

with Φ, a diagonal matrix having on its diagonals the phase-shifts


between the two arrays for the P sources:

j 2π
λ 1
T
k δ
e 1

Φ= ⋱ .
j 2π T
k δ
λp P
e

In the expression of Φ, λp denotes the p th source wavelength.


Thus, the DOA estimation problem comes down at estimating the
Φ matrix. To this end, the covariance matrix Ryy of the entire
Fig. 9 Acquisition set-up for ESPRIT
array:

700 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
y1(t) A b1(t) 4.1 CPD-based array processing
y(t) = = x(t) +
y2(t) AΦ b2(t) To overcome these drawbacks, a DOA estimation approach was
introduced in [5], capable of handling N arbitrarily displaced
is first computed as: subarray, as illustrated in Fig. 10.
It is also the first time that a tensor-based algorithm is proposed
1
K
for the DOA estimation problem. If y1(t), …, yN (t) are the outputs
K k∑
Ryy = E y(t)yH(t) ≃ y(tk)yH(tk) .
of the N subarrays and using notations similar to those used for
=1
ESPRIT, we have:
T
Then, we use the fact that the first P eigenvectors E1T, E2T of Ryy P
span the same subspace as the source steering vectors, i.e. y1(t) = ∑ a(kp)sp(t) + b (t) = Ax(t) + b (t),
1 1
p=1
E1 A AT P
j 2π kT δ
= T= , (74) y2(t) = ∑ a(kp)e λ p 2
p sp(t) + b2(t) = AΦ2 x(t) + b2(t),
E2 AΦ AΦT
p=1

where T is a non-singular matrix. As ⋮


span{ A} = span{ E1} = span{ E2}, there exist a non-singular matrix P
j 2π kT δ
Ψ such that yN (t) = ∑ a(kp)e λ p N
p sp(t) + b2(t) = AΦN x(t) + bN (t),
p=1
E1Ψ = E2 . (75)
where
From (74) and (75), it can be shown, after some algebraic
j 2π T
manipulations (see e.g. [45]), that Φ can be directly obtained by e
k δ
λ 1 n
1

the diagonalisation of the matrix Ψ = E1† E2. However, this least- Φn = ⋱ , n = 2, . . . , N .


squares (LS) estimate Ψ is biased because it assumes implicitly
j 2π T
that E1 is perfectly known and the error is to be attributed only to
k δ
λ P n
e P

E2. In practice, E1 and E2 are equally ‘noisy’ as they are both


estimates of span{ A}. A more appropriate criterion that takes into If K snapshots are considered, the output of the entire array can be
account noise on both E1 and E2 is the total least-squares (TLS) expressed as the (NL × K) matrix Y:
criterion. It can be formulated as finding two minimum Frobenius
norm residual matrices R1 and R2, and a matrix Ψ satisfying Y1 A1 B1
E1 + R1 Ψ = E2 + R2. If we denote B = E1 + R1, the TLS-ESPRIT Y2 A1Φ2 T B2
problem can be expressed as the following minimisation problem: Y= = S + = (A ⊙ D)ST + B, (77)
⋮ ⋮ ⋮
E1 B 2 YN A1ΦN BN
min ∥ − ∥ . (76)
Ψ, B E2 BΨ F where S = s1, …, sP is a K×P source matrix,
A = a(k1), …, a(kP) is the (L × P) steering vector matrix and
The solution of (77) is obtained by the eigenvalue decomposition
of the 2P × 2P matrix E1 E2 H E1 E2 .The ESPRIT algorithm
diag{IP}T
presents several advantages over the ‘classical’ DOA estimation
methods such as the beamforming or MUSIC [46, 47]. For diag{Φ2}T
example, it requires no calibration step and parameters estimates D = [d(k1), …, d(kP)] = is a (N × P)

are obtained directly via two well-conditioned eigenproblems.
Thus ESPRIT avoids the parameter grid search, inherent to diag{ΦN }T
MUSIC-like algorithms. However, in order to use ESPRIT one
needs to estimate the data covariance matrix, which can be a matrix regrouping on its rows the diagonal elements of Φn, B is a
difficult task in the case of highly correlated sources. Another noise matrix. In the light of the concepts introduced in Sections
major drawback of ESPRIT is the fact that it cannot handle more 3.1.2 and 3.1.3, (77) expresses a noisy CP decomposition of the
than two displaced subarrays. sensor array output. Using the tensor formalism, (77) can be
expressed as a (L × K × N) tensor Y:

P
Y= ∑ ap ⊗ sp ⊗ d p + ℬ = A, S, D; ℐ3, P + ℬ . (78)
p=1

The tensor form (78) clearly highlights the three diversities/


invariances of the array output: sensors – time – subarrays. To
estimate the loading matrices A, S, D, and consequently, the source
DOAs, an approximate CP decomposition of the data tensor Y
must be performed, using one of the many existing algorithms.
Unlike the previously presented methods, the CP approach applies
directly to the ‘raw’ data and does not require the estimation of the
second-order statistics of the array output.
In [25], the ill-posedness of the tensor approximation problem
induced by (78) is solved thanks to a constraint imposing that
sources must be different, in the sense that either they should stem
from different directions (e.g. if they are fully correlated like
multipaths), or they should not be fully correlated, or both.
In [48], an approach that generalises the results given in [5] to
Fig. 10 Acquisition set-up for the CP approach an array that presents an arbitrary number of spatial invariances
was proposed. Consider a subarray composed of L1 isotropic
IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 701
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
where n(t) is a complex-valued zero-mean additive white noise.
Let us assume that we have at our disposal K snapshots at time
instants, t1, t2, …, tK . Define the following Ln × P matrices:

A1 = a1(k1), …, a1(kP)
(83)

AN = aN (k1), …, aN (kP) , (84)

and the K × P matrix:

s1(t1) s2(t1) … sP(t1)


s1(t2) s2(t2) … sP(t2)
S= = s1, s2, …, sP . (85)
⋮ ⋮ ⋱ ⋮
s1(tK ) s2(tK ) … sP(tK )

The collection of K snapshots of the array can then be organised


Fig. 11 Multi-scale planar array with three hierarchical levels into a L × K data matrix as
identical sensors indexed by l1 = 1, …, L1. Consider then, L2 Z = z(t1), …, z(tK ) = A1 ⊙ … ⊙ AN ST + N, (86)
identical replicas of this subarray, spatially translated to arbitrary
locations. The L2 distinct copies of the original subarray, indexed
where N is a (L × K) complex-valued matrix, modelling the noise
by l2 = 1, …, L2, can now be seen as subarrays that together
on the entire array for all K temporal snapshots. Equation (86)
constitute a larger (higher-level) array. This proposed array reveals a (N + 1) -dimensional CPD structure. In the case where
structure can be further generalised by considering an additional
only one time-sample is available, i.e. matrix S is a 1 × P vector,
level, composed of L3 translated replicas of the previous sensor
the data model given by (86) becomes
subarrays, indexed by l3 = 1, …, L3. Let us generalise this scheme
to a total of N such hierarchical levels, with the ‘highest’ level z = (A1 ⊙ … ⊙ AN )s + n, (87)
consisting of LN subarrays indexed by lN = 1, …, LN . It is worth
noting that two different subarrays at a given level n need not be
disjoint, i.e. they may have in common subarrays/sensors of the with z = z(t1), s = s(t1) = (S(1, : ))T and n = N(: , 1). In the
previous level n − 1 . However, if all subarrays at each level are definitions above, we used the Matlab notations for columns and
disjoint, then the entire array will contain a total number of rows selection operators. Equation (87) is a vectorised
L = L1L2…LN identical sensors. Fig. 11 illustrates a three-level representation of a N-dimensional CP data model. It is worth
noting that if only one snapshot of the array is available, the N + 1
array of co-planar sensors.
CP model degenerates into an N -dimensional one. A discussion on
Consider first a narrow-band plane wave impinging on the array
the identifiability of these CPD models, for different scenarios can
as described in the beginning of Section 4. Let us denote by al1l2…lN
be found in [48].
its phase factor at the sensor indexed by l1, l2, …, lN at the N various For this multi-scale invariance array, a two stages DOA
T
levels of the array. Define dl(nn) = xl(nn) yl(nn) zl(nn) , with n = 1, …, N. estimation procedure was also proposed in [48]. The first stage
With the notation introduced above, the spatial phase factor is estimates the N steering vectors an(kp) (n = 1, …, N) for each of
given by: the P sources (p = 1, …, P), by exploiting the CPD structure (87)
of the collected data. The second stage estimates the sources’
N N direction-cosines kp, p = 1, …, P from the steering vectors
2π 2π T (n)
al1l2…lN (k) = exp j
λ ∑ kTdln ( )
n
= ∏ exp j
λ
k dln . (79) obtained at the previous stage. To this end, the DOA estimation can
n=1 n=1 be formulated as an optimisation problem, and a sequential
procedure that exploits all the available information from the
Thus, the array manifold for the entire sensor-array is source's steering vectors encompassing all scale levels, is adopted.
Define the following cost functions:
a(k) = a1(k) ⊠ ⋯ ⊠ aN (k), (80)
p
Jn(kp) = ∥ a^ n − an(kp) ∥22 , with n = 1, …, N, (88)
with
p
T (n) where a^ n denotes the estimated steering vector at the nth level for
e j(2π / λ)k d1
the pth source. Estimating the DOAs for the pth source comes
an(k) = ⋮ (81) down to minimising the following criterion:
(n)
j(2π / λ)kTdLn
e N
ℐN (kp) = ∑ Jn(kp) . (89)
an Ln × 1 vector, ∀n = 1, …, N. n=1
Next, consider P narrow-band plane-waves, having traveled
through a non-conductive homogeneous isotropic medium, This function is non-convex and highly non-linear with respect to
impinging upon the array from directions kp, with p = 1, …, P. the direction-cosines; hence, a direct local optimisation procedure
Denote by sp(t) the time signal emitted by the pth narrow-band would fail in most cases. In [48], a sequential strategy was adopted
source. Then, the output at time t of the entire sensor-array can be to minimise ℐN kp , progressing from one level to the next higher
expressed as an L × 1 vector, level, using iterative refinement of the direction-cosine estimates
within each level. The method is based on the fact that, when
P noise-free, the N cost-functions in (89) have the same global
z(t) = ∑ (a (kp) ⊠ ⋯ ⊠ aN (kp))sp(t) + n(t),
1 (82) minimum.
p=1 Assume that the level-1 subarrays’ inter-sensor separations do
not exceed half a wavelength. This assumption is essential to

702 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
obtaining a set of high-variance but unambiguous direction-cosine computations over the cores [61]. In the tensor network framework,
estimates. On the contrary, to achieve a practical advantage, it is Hierarchical/tree Tucker [29, 64] and tensor train (TT) [28] are two
important that the spatial displacement between any two subarrays popular representations of a high-order tensor into a graph-
of the highest level exceeds λ/2, where λ is the wavelength. This connected low-order (at most 3) tensors. In this section, we focus
will produce estimates of lower variance but with cyclic ambiguity our effort on the TT formalism for its simplicity and compactness
for the same set of direction-cosines. On the other hand, under the in terms of storage cost. Unlike the hierarchical Tucker model, TT
first assumption, the J1(kp) function is unimodal on the support is exploited in many practical and important contexts as, for
region of the DOAs. Therefore, any local optimisation procedure instance, tensor completion [65], blind source separation [66], and
should converge towards the global minimum for the criterion. machine learning [4], to mention a few. In the context of the MHR
Thus, we obtain another set of estimates, now of high-variance but problem, this strategy has at least two advantages. First, it is well-
with no cyclic ambiguity, for the DOAs, to be denoted by k∗p, 1 with known that the convergence of the ALSs algorithm becomes more
p = 1, …, P. These estimates will subsequently be used, in a and more difficult when the order increases [67–69]. To deal with
second step, as the initial point for the minimisation of this problem, applying ALS on lower-order tensors is preferable.
ℐ2 kp = J1 kp + J2 kp . As no assumption is made on the The second argument is to exploit some latent coupling properties
between the cores [4, 70] to propose new efficient estimators.
distances between the level-2 subarrays, ℐ2 kp may present more The maximum likelihood estimator (MLE) [71, 72] is the
than one local minimum. Hence, a good initial estimate is crucial optimal choice from an estimation point of view since it is
for the optimisation procedure. The estimates obtained by the statistically efficient, i.e. its mean squared error (MSE) reaches the
minimisation of ℐ2 kp , denoted by k∗p, 2, are then used for the Cramér–Rao bound (CRB) [73] in the presence of noise. The main
minimisation of ℐ3 kp = ∑n = 1 Jn kp , and so on, until the final
3
drawback of the MLE is its prohibitive complexity cost. This
estimates are obtained by the minimisation of ℐN kp . The limitation is particularly severe in the context of a high-order data
proposed algorithm can be summarised as follows: tensor. To overcome this problem, several low-complexity methods
First stage: can be found in the literature. These methods may not reach,
• Estimate A1, …, AN by CP decomposition of the data Z or z sometimes, the CRB, but they provide a significant gain in terms of
the computational cost compared to the MLE. There are essentially
[see (86) or (87)].
two main families of methods. The first one is based on the
Second stage:
factorisation of the data to estimate the well-known signal/noise
• For p = 1, …, P and for n = 1, …, N, compute
subspace such as the estimation of signal parameters via rotational
n
invariance techniques (ESPRIT) [45], the ND-ESPRIT [74], the
(90) improved multidimensional folding technique [75], and the CP-
k∗p, n = argmin ℐn kp = argmin
kp kp
∑ Ji kp .
VDM [76]. The second one is based on the uniqueness property of
i=1
the CPD. Indeed, factorising the data tensor thanks to the ALS
• Output: The estimated parameters for the P sources: algorithm [77] allows identification of the unknown parameters by
^ Vandermonde-based rectification of the factor matrices.
k p = u^ p, v^ p, w^ p = k∗p, N with p = 1, …, P.
Several CPD-based DOA estimation approaches, based on other
diversity schemes have been proposed in the literature. In [49], 5.1 Generalised Vandermonde canonical polyadic
polarisation of electromagnetic waves was exploited as an decomposition
additional diversity, by using vector-sensor arrays capable of The MH model assumes that the measurements can be modeled as
capturing the six components of the electromagnetic field. A the superposition of R undamped exponentials sampled on a Q -
coupled CPD approach for DOA estimation, using the multiple dimensional grid according to [6]
baselines in sparse arrays was introduced in [50]. High-order
statistics (cumulants) were also used to create diversity in CPD R Q
based direction-finding algorithms [51]. [X]i1iQ = ∑ αr ∏ zri q q−1
, , 1 ≤ iq ≤ Iq (91)
While CPD remains the most commonly used decomposition in r=1 q=1
tensor-based array processing, other tensor decompositions made
their way in. For example, the HOSVD has been used to develop in which the rth complex amplitude is denoted by αr and the pole is
multidimensional versions of the popular ESPRIT and MUSIC defined by zr, q = e jωr, q where ωr, q is the rth angular-frequency
algorithms (see e.g. [52, 53]). T
along the qth dimension, and we have zq = z1, q z2, q … zR, q .
Note that the tensor X is expressed as the linear combination of M
5 Tensor-based MHR rank-1 tensors, each of size I1 × × IQ (the size of the grid), and
MHR [6, 54] is a classical signal processing problem that has follows a generalised Vandermonde CPD [78]:
found several applications in spectroscopy [55], radar
communications [56], sensor array processing [5, 57], to mention a X = A ∙ V1 ∙ … ∙ VQ (92)
few. The multidimensional harmonic (MH) model can be viewed 1 2 Q
as the tensor-based generalisation of the one-dimensional harmonic
one, resulting from the sampling process over a multidimensional where A is a R × ⋯ × R diagonal tensor with [A]r, , r = αr and
regular grid. As a consequence, the Q-dimensional harmonic model
needs the estimation of a large number QR of angular-frequencies Vq = v(z1, q) ⋯ v(zR, q)
of interest. We can easily note that the number of unknown
parameters and the order of the associated data tensor grow with Q. is a Iq × R rank- R Vandermonde matrix, where
Moreover, it is likely that the joint exploitation of multi-diversity/
modality sensing technologies for data fusion [58–60] further I −1 T
v(zr, q) = 1 zr, q zr2, q … zrq, q .
increases the data tensor order. This trend is usually called the
‘curse of dimensionality’ [61–63] and the challenge here is to
reformulate a high-order tensor as a set of low-order tensors over a We define a noisy MH tensor model of order Q as:
graph. In this context, we observe an increasing interest for the
tensor network theory (see [61] and references therein). Tensor Y = X + σℰ, (93)
network provides a useful and rigorous graphical representation of
a high-order tensor into a factor graph where the nodes are low- where σℰ is the noise tensor, σ is a positive real scalar, and each
order tensors, called cores, and the edges encode their entry [ℰ]i1iQ follows an i.i.d circular Gaussian distribution
dependencies, i.e. their common dimensions, often called ‘rank’. In CN(0, 1), and X has a canonical rank equal to R. The reader may
addition, tensor network allows to perform scalable/distributed
IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 703
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
the structure (Fig. 12). The most intuitive way to exploit the
1
Vandermonde structure is called column-averaging. Let ω = ∠zi
i
where ∠ stands for the angle function. Define the sets

f
J= v= : f ∈ ℂI (94)
[ f ]1


1 1
(95)
ℓ i∑
Aℓ = v z = e jω̄ : ω̄ = ∠[ f ]i + 1
=1
i

where 1 ≤ ℓ ≤ I − 2. This method exploits the Vandermonde


structure in a heuristic way. So, the rectified strategy is to consider
h
the iterated vector f (h) = πAℓπJ ( f ) . Another straightforward
method is the periodogram maximisation. Under Gaussian noise
and for a single tone αrzrq, q, the MLE is optimal and is given by the
i

location of the maximal peak of the Fourier-periodogram [8, 82].


To increase the precision of the estimation, it is standard to use the
well-known zero-padding technique at the price of an increase in
computational cost. In [83], two strategies for Vandermonde-based
rectification of the CP factors are provides but the RecALS
principle is generic and flexible in the sense that any
Vandermonde-based rectification methods can be used.

5.3 VTT-RecALS
In this section, the JIRAFE algorithm and the RecALS method are
associated to solve the MHR problem. The VTT-RecALS estimator
is based on the JIRAFE principle which is composed of two main
steps.
(i) The first one is the computation of the TTD of the initial
tensor. By doing this, the initial Q-order tensor is broken down into
Q graph-connected third-order tensors, called TT-cores. This
dimensionality reduction is an efficient way to mitigate the ‘curse
Fig. 12 Algorithm 4 Rectified ALS (RecALS) of dimensionality’. To reach this goal, the TT-SVD [28] is used as
a first step.
be referred to [74] for the case of damped signals, which is not (ii) The second step is dedicated to the factorisation of the TT-
addressed in this paper. cores. Recall the main result given by Theorem 2, i.e. if the initial
tensor follows a Q-order CPD of rank R, then the TT-cores for
5.2 Algorithms 2 ≤ q ≤ Q − 1 follow coupled third-order CPD of rank R.
Consequently, the JIRAFE minimises the following criterion over
5.2.1 Vandermonde rectification of the ALS algorithm: Limit the physical quantities {V1, …, VQ} and over the latent quantities
of the ALS algorithm for structured CPD: The CPD of any order- {M1, …, MQ − 1}:
Q rank-R tensor X involves the estimation of Q factors Vq of size
Iq × R. As pointed out above, in the context of the MD-harmonic 2 2
C= G1 − V1 AM1−1 + GQ − MQ − 1VQT (96)
model, the factors Vq of the CPD are Vandermonde matrices. F F

Consider, Yq, the qth mode unfolding [9] of tensor Y, at the kth 2
Q−1
iteration with 1 ≤ k ≤ M, M denoting the maximal number of +∑ Gq − ℐ3, R ∙ Mq − 1 ∙ Vq ∙ Mq− T , (97)
iterations. The ALS algorithm solves alternatively for each of the Q q=2 1 2 3
F
dimensions the minimisation problem [23, 71]:
2
minVq Yq − VqSq where where A is a R × R diagonal matrix with [A]r, r = αr.
SqT= VQ ⊙ … ⊙ Vq + 1 ⊙ Vq − 1 ⊙ … ⊙ V1. It aims at approximating The above cost function is the sum of coupled LS criteria. The
tensor Y by a tensor of rank R, hopefully close to X. The LS aim of JIRAFE is to recover the original tensor factors using only
solution conditionally to matrix Sq is given by Vq = YqSq† where † the third-order tensors Gq. Consequently, the JIRAFE approach
stands for the pseudo-inverse. Now, remark that there is no reason adopts a straightforward sequential methodology, described in
that the above LS criterion promotes the Vandermonde structure in Fig. 13, to minimise the cost function C.
the estimated factors in the presence of noise. In other words, We present in Algorithm 5, the pseudo-code of the VTT-
ignoring the structure in the CPD leads to estimate an excessive RecALS algorithm, where RecALS3 denotes the RecALS applied
number of free parameters. This mismatched model dramatically to a third-order tensor, while RecALS2 denotes the RecALS applied
decreases the estimation performance [79]. Hence there is a need to to a third-order tensor using the knowledge of one factor. The VTT-
rectify the ALS algorithm to take into account the factor structure. RecALS algorithm is actually divided into two parts (Fig. 14). The
Rectified ALS (RecALS): The RecALS algorithm belongs to the first part is dedicated to dimensionality reduction, i.e. breaking the
family of lift-and-project algorithms [80, 81]. The optional lift step dimensionality of the high Q-order tensor into a set of third-order
computes a low rank approximation and the projection step tensors using Theorem 2. The second part is dedicated to the
performs a rectification toward the desired structure. The RecALS factors retrieval from the TT-cores using the RecALS algorithm
algorithm is based on iterated projections and split LS criteria. Its presented in the previous section. It is worth noting that the factors
algorithmic description is provided in Algorithm 4 for Q = 3. We Vq are estimated up to a trivial (common) permutation ambiguity
insist that several iterations in the while loops are necessary, since [84]. As noted in [37], since all the factors are estimated up to a
restoring the structure generally increases the rank, and computing unique column permutation matrix, the estimated angular-
the low-rank approximation via truncated SVD generally destroys frequencies are automatically paired.

704 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
• to derive closed-form receivers under a priori knowledge of
some transmitted symbols (semi-blind systems);
• to deal with point-to-point systems as well as multi-hop systems
with relays.

A possible classification of MIMO wireless communication


systems can be made according to:

• the considered channel access and multiplexing technology like


CDMA (code-division multiple access), TDMA (time-division
multiple access), FDMA (frequency-division multiplexing
access), OFDMA (orthogonal frequency-division multiplexing
Fig. 13 VTT-RecALS representation access), or yet hybrid technologies like CDMA–OFDM or
TDMA–FDMA;
• the assumptions on the channel: flat fading versus frequency-
selective fading channel, according to whether all the frequency
components of the transmitted signals are attenuated by the
same fading or not; in this last case, the frequency dependency
of the channel coefficients leads to a third-order MIMO channel
tensor;
• the consideration or not of a sparsely scattering environment
characterised by a multipath assumption with DoD (direction-of-
departure) and DoA (direction-of-arrival) angles;
• the consideration or not of resource allocation under the form of
matrix or tensor allocation;
• the type of coding (matrix/tensor), and the diversities taken into
account (ST, STF codings);
• the type of communication system (supervised versus semi-
blind, depending on the use or not of a training sequence);
• the type of receiver (iterative versus closed-form);
• the type of optimisation algorithm used at the receiver (ALS,
Levenberg–Marquardt...);
Fig. 14 Algorithm 5 VTT-RecALS • the type of communication: half-duplex (HD) versus full-duplex
(FD) depending on the possibility or not of a simultaneous
communication between two users, meaning that the
6 Tensor-based MIMO wireless communication
communication can be in only one direction or in both
systems directions, respectively;
As it is now well known, the use of multiple transmit and receive • the existence or not of relays corresponding to multi-hop and
antennas, corresponding to a MIMO channel, allows to single-hop communication systems, respectively;
significantly improve the performance of wireless communication • the relaying protocol used at the relays (amplify-and-forward
systems [85]. Such MIMO systems exploit space diversity to (AF) versus decode-and-forward (DF) protocols).
combat channel fading. Each information symbol is transmitted
through different paths between the transmitter and the receiver, Note that the DF protocol means that the signals received at the
each path being characterised by an attenuation modeled by means relay are decoded and then re-encoded before to be forwarded by
of a fading coefficient of the channel. These channel coefficients the relay, whereas in the case of the AF protocol, the received
are generally assumed to be independent and identically distributed signals are only amplified and forwarded by the relay. When two
(i.i.d.), which corresponds to an i.i.d. Rayleigh flat fading channel. end nodes are communicating with each other through a relay, via
In the presence of a slow fading environment, the channel bidirectionnal links, one can distinguish one-way and two-way
coefficients are also assumed to be constant during P time slots, the relaying depending on the way the data streams are exchanged
duration of transmission of N data streams. between the end nodes via the relay. With two-way relaying, the
Another way to improve the performance consists in using end nodes transmit simultaneously their data to the relay, and then
space–time (ST) or space–time–frequency (STF) codings to get the relay combines this information before broadcasting them to
space, time and frequency diversities, corresponding to the end nodes, which allows to increase the system throughput (in
redundancies of the information symbols in each of these domains bits/channel use).
[33, 86, 87]. One objective of MIMO communication systems is to A specificity of tensor-based communication systems lies in the
maximise the combined diversity gain while ensuring the best development of new tensor models through the design of new
transmission rate possible, which corresponds to the fundamental systems. For instance, in the case of point-to-point systems,
diversity-rate tradeoff. Since the pioneering work [7], wireless CONFAC (constrained factors) [88], generalised PARAFAC-
communications have been a privileged field of application for Tucker (PARATUCK) [86], and nested PARAFAC models [34]
tensors. Indeed, tensor-based approaches in the context of MIMO were proposed. Most of these models can be viewed as constrained
communication systems allow: PARAFAC decompositions. See [11, 89, 90] for overviews of such
models and corresponding point-to-point MIMO systems. In the
• to take simultaneously into account different diversities like context of cooperative or relay-assisted communications, several
space, time, frequency or polarisation, leading to high order papers have proposed tensor-based models and algorithms for the
tensor models for the signals received at destination; supervised channel estimation problem [91, 92], considering two-
• to design tensor coding like tensor space–time (TST) or tensor hop AF relays and two-way MIMO relaying [93]. Tensor-based
space–time–frequency (TSTF) codings; semi-blind receivers for the uplink multiuser cooperative scenario
• to develop semi-blind receivers for jointly estimating the were first proposed in [94, 95]. For the two-hop MIMO AF
information symbols and the channel, by exploiting uniqueness relaying case, tensor modeling was exploited in several papers to
properties of tensor models, which is an advantage of tensor- derive semi-blind joint channel and symbol estimation algorithms
based systems over matrix-based ones; [35, 96, 97]. More recently, tensor-based receivers have been

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 705
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
developed for multi-hop relaying systems, by exploiting the nested U = C ∙ S ∈ ℂM × N × P (102)
Tucker and coupled nested Tucker models [31, 98]. 2
In this section, we make a brief presentation of three point-to-
point systems and of one relaying system to illustrate the tensor- or equivalently:
based approach in the context of wireless communication systems.
R
um, n, p = ∑ cm r psn r, (103)
6.1 Khatri–Rao ST coding r=1
, , ,

Let us consider a MIMO wireless system with M transmit and K


receive antennas. The information symbols to be transmitted by the which can be viewed as a linear combination of the R information
M transmit antennas define a symbol matrix S ∈ ℂN × M , composed symbols belonging to the nth data stream, with the coding
coefficients cm, r, p as coefficients of the linear combination.
of N data streams, each one containing M symbols. These symbols
After transmitting the coded signals from the M transmit
are coded using a ST coding matrix C ∈ ℂP × M such that the coded
symbols to be transmitted are given by: antennas via the channel H ∈ ℂK × M , the signals received at
destination satisfy the following equation:
U = C ⊙ S ∈ ℂPN × M , (98)
X = U ∙ H + ℬ ∈ ℂK × N × P
1
or equivalently: (104)
=C∙H∙S+ℬ
1 2
u( p − 1)N + n, m = cp, msn, m . (99)
or in scalar form:
The coding matrix C introduces a time diversity in the sense that
every symbol sn, m is duplicated P times, with P corresponding to M
the number of time slots, and the channel being assumed to be xk, n, p = ∑ um n phk m + bk n p
, , , , ,

constant during the transmission, i.e. for p ∈ [1, P]. The coding m=1
(105)
(98) can be viewed as a simplified version of the Khatri–Rao M R
space–time (KRST) coding proposed in [99]. = ∑ ∑ cm r phk msn r + bk n p,
, , , , , ,

The MIMO channel is assumed to be i.i.d. Rayleigh flat fading m=1r=1

and defined as H ∈ ℂK × M , where hk, m is the complex fading


which defines a third-order tensor satisfying a Tucker-(2,3) model
coefficient between the mth transmit antenna and the kth receive [11], with the coding tensor C as core tensor, and the factor
antenna. The transmitter transmits the coded signals via the matrices (H, S, IP), where IP is the identity matrix of order P. One
communication channel, so that the signal received by the kth
factor matrix being an identity matrix, such a Tucker model is
receive antenna, during the pth time slot, and associated with the
unique up to two non-singular matrices (ΔH, ΔS). If the coding
nth data stream, can be written as:
tensor is assumed to be known at the receiver, the model ambiguity
M becomes a scalar factor related to a Kronecker product. This scalar
xk, n, p = ∑ hk mu p
, ( − 1) N + n, m + bk, n, p ambiguity can be eliminated assuming the a priori knowledge of
m=1 only one symbol, leading to a semi-blind Kronecker receiver [90].
M
(100) Then, the channel and symbol matrices can be jointly estimated
= ∑ hk, msn, mcp, m + bk, n, p . using the Kronecker factorisation (KF) algorithm (see Section
m=1 3.2.5), after the LS estimation of the Kronecker product H ⊠ S as:

These received signals form a third-order tensor X ∈ ℂK × N × P,



H ⊠ S = XKN × PCMR × P, (106)
which satisfies a CPD model, with rank M and factor matrices
(H, S, C), and ℬ ∈ ℂK × N × P represents the additive noise tensor. where CMR × P denotes a tall mode-3 matrix unfolding of the coding
These matrices can be estimated using the tri-ALS algorithm. tensor. Choosing this coding tensor such that its matrix unfolding
Assuming the coding matrix known at the receiver, it is possible to CMR × P is row-orthonormal allows to simplify the right-inverse as
use the bi-ALS algorithm for jointly estimating the channel and †
CMR ∗ ∗
× P = CP × MR[CMR × P(CP × MR]
−1
= CP∗ × MR, with the necessary
symbol matrices. Another solution consists in applying the Khatri– condition MR ≤ P for uniqueness of this inverse.
Rao factorisation (KRF) algorithm [43] after a LS estimation of the Comparing the two presented MIMO systems which are
Khatri–Rao product H ⊙ S as: simplified versions of the KRST and TST coding-based systems
([99] and [33], respectively), one can conclude that both systems
H ⊙ S = XKN × P(CT)†, (101) lead to a third-order tensor for the received signals, with the same
dimensions. This tensor satisfies a PARAFAC model, in the first
where ( . )† denotes the matrix right inverse, and XKN × P represents case, while it satisfies a Tucker-(2,3) model in the second case,
a tall mode-3 matrix unfolding of the received signals tensor which induces different ambiguities. Another important difference
is that with the TST coding, the number (R) of symbols per data
X ∈ ℂK × N × P (see Section 3.2.6). It is worth noting that choosing a
stream can be different from the number (M) of transmit antennas,
column-orthonormal coding matrix allows to simplify the right-
which is not the case with the KRST coding.
inverse of CT as (CT)† = C∗(CTC∗)−1 = C∗, with the necessary In a practical situation where the coding used at the transmitter
condition M ≤ P for uniqueness of this inverse. is known at the receiver, one can draw the supplementary
conclusion that both systems offer the possibility of jointly and
6.2 Tensor space–time coding (TSTC) semi-blindly estimating the channel and symbol matrices by means
of a closed-form receiver. However, in order to eliminate model
We now present another MIMO system which can be viewed as a
ambiguities, the Kronecker factorisation algorithm used with TST
simplified version of the TST one [33] in the sense that no resource
coding presents the advantage over the Khatri–Rao factorisation
allocation matrix is considered. A TST coding C ∈ ℂM × R × P is algorithm used with KRST coding, to require the a priori
used to code the symbol matrix S ∈ ℂN × R, composed of N data knowledge of only one symbol for the Kronecker model while the
streams, each one containing R symbols. a priori knowledge of M symbols is needed for the PARAFAC
The coded signals to be transmitted are given by: model.

706 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
where GRMF × JPF is a matrix unfolding of the code-allocation tensor
defined in (108). By properly designing this tensor to ensure its
right-invertibility, the channel and symbol matrices are estimated in
closed form from a LS estimation of the Kronecker product

S ⊠ HK × MF = XNK × JPFGRMF × JPF . (111)

The use of a pilot-assisted channel estimation and of a semi-blind


Fig. 15 Block diagram of a one-way two-hop MIMO relay system ALSs-based receiver are discussed in [86].

6.3 TSTF coding 6.4 MIMO relaying systems with tensor coding
A more general and flexible tensor-based transceiver scheme for The benefits of TSTC have been extended to MIMO relaying
space, time and frequency transmit signalling was proposed in [86], systems in [31], where semi-blind receivers have been derived for
referred to as TSTF coding. By relying on a new class of tensor the joint estimation of the source-relay and relay-destination
models, namely, the generalised PARATUCK- N1, N and the channels, in addition to symbol estimation. Consider a one-way
Tucker- N1, N models, the TSTF system combines a fifth-order two-hop MIMO relaying communication system shown in Fig. 15,
coding tensor with a fourth-order allocation tensor. It can also be where M, and K denote the numbers of antennas at the source S ,
seen as a generalisation of three previous tensor-based and destination D nodes, respectively, while QR and QT represent
ST/TST/STF coding schemes, offering new performance/ the numbers of receive and transmit antennas at the relay,
complexity tradeoffs and space, time and frequency allocation
respectively. The source–relay channel H(SR) ∈ ℂQR × M and the
flexibility.
Herein, the transmission time is decomposed into P time-slots relay–destination channel H(RD) ∈ ℂK × QT are assumed to be flat
(data blocks) of N symbol periods, each one being composed of J fading and quasi-static, i.e. constant during the transmission
chips. At each symbol period n of the pth block, the transceiver protocol which is divided into two consecutive phases. Assume
transmits a linear combination of the nth symbols of certain data that the direct link between the source and destination is absent or
streams, using a set of transmit antennas and of sub-carriers. The negligible. During the first hop, the transmission is composed of N
coding is carried out by means of a fifth-order code tensor time-blocks associated with N symbol periods, R data streams
W ∈ ℂM × R × F × P × J whose dimensions are the numbers of transmit being transmitted per time-block. The transmitted symbol matrix
antennas M , data streams R , sub-carriers F , time blocks P , S ∈ ℂN × R contains R data streams composed of N symbols each.
During the time-block n, each antenna m of the source node
and chips J . An allocation tensor C ∈ ℂM × R × F × P determines the
transmits a combination of R information symbols
transmit antennas and the sub-carriers that are used, as well as the
sn, r, r = 1, . . . , R, to the relay, after a TSTC [31] by means of a
data streams transmitted in each block p. As an example,
cm, r, f , p = 1 means that the data stream r is transmitted using the third-order tensor C(S) ∈ ℂM × P × R, which introduces ST
transmit antenna m, with the sub-carrier f, during the time block p. redundancies, since each symbol sn, r is repeated P times over the M
The TSTF-coded signals to be transmitted are given by: transmit antennas, P being the source code length. During the
second hop, the source remains silent and the relay uses a second
U = G ∙ S, with G = W ⋆ C, (107) three dimensional coding C(R) ∈ ℂQT × J × QR before forwarding the
1 {m, r, f , p}
QR received signals to the destination using QT transmit antennas.
or, alternatively, in scalar form Not that this second tensor coding consists in repeating the signals
received at the relay J times, over the QT transmit antennas, J being
R the relay code length. The tensor T ∈ ℂQT × J × P × N of signals
um, n, f , p, j = ∑ wm r f p jsn rcm r f p,
, , , , , , , , (108) transmitted by the relay is defined as
r=1
QR M R
where G ∈ ℂM × R × F × P × J is the effective code-allocation tensor, tqT, j, p, n = ∑ ∑ ∑ cqR j q hqSRmcmS p rsn r .
( ) (
T, , R R,
) ( )
, , , (112)
given by the Hadamard product between the code tensor qR = 1 m = 1 r = 1

W ∈ ℂM × R × F × P × J and the allocation tensor C ∈ ℝM × R × F × P


along their common modes, i.e. After transmission through the relay–destination channel, the
noiseless signals received by K antennas at destination satisfies the
gm, r, f , p, j = wm, r, f , p, jcm, r, f , p . following equation:
QT
Assuming a flat Rayleigh fading propagation channel, the discrete-
time baseband-equivalent model for the signal received at the kth
xk, j, p, n = ∑ hkRDq tq
(
, T
)
T, j , p , n
qT = 1
receive antenna during the jth chip period of the nth symbol period (113)
of the pth block, and associated with the fth sub-carrier, is given QT QR M R
(RD) (R) (SR) (S)
by: = ∑ ∑ ∑ ∑h c h c
k, qT qT, j, qR qR, m m, p, r sn, r .
qT = 1 qR = 1 m = 1 r = 1
M R
xk, n, f , p, j = ∑ ∑ hk m f sn rwm r f p jcm r f p + bk n f p j .
, , , , , , , , , , , , , ,
Comparing (113) with (31), we can conclude that the tensor
m=1r=1 X ∈ ℂK × J × P × N of noiseless signals received at destination
satisfies a NTT(4) model. Fig. 16 illustrates the NTT(4) for the
The above equation can be written as the following generalised noiseless signal tensor X received at the destination. We have to
Tucker-(2,5) model [11] note that the four dimensions of the tensor X are associated with
four signal diversities (space k, relay code j, source code p, time n).
X = G ∙ ℋ ∙ S. (109) A special case of this model is the one presented in [35, 96],
1 2
where the coding tensors C(S) and C(R) have a special structure
Joint channel and symbol estimation is carried out by exploiting where their frontal slices are diagonal matrices, with M = R and
the following unfolding of the received signal tensor: QR = QT. Physically, this means that the source and the relay
transmits a single data stream, in contrast to the NTT(4) model in
XNK × JPF = S ⊠ HK × MF GRMF × JPF, (110) (114) where R and QR signals are combined at the source and relay,

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 707
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
[8] Nion, D., Sidiropoulos, N.D.: ‘Tensor algebra and multidimensional harmonic
retrieval in signal processing for mimo radar’, IEEE Trans. Signal Process.,
2010, 58, (11), pp. 5693–5705
[9] Comon, P.: ‘Tensors: a brief introduction’, IEEE Signal. Proc. Mag., 2014, 31,
(3), pp. 44–53, special issue on BSS
[10] De Lathauwer, L., Moor, B.D., Vandewalle, J.: ‘A multilinear singular value
decomposition’, SIAM J. Matrix Anal. Appl., 2000, 21, (4), pp. 1253–1278
[11] Favier, G., de Almeida, A.: ‘Overview of constrained PARAFAC models’,
Fig. 16 NTT(4) model for a one-way two-hop MIMO relaying system with EURASIP J. Adv. Signal Process., 2014, 5, pp. 1–41
tensor coding [12] Hitchcock, F.L.: ‘The expression of a tensor or a polyadic as a sum of
products’, J. Math Phys., 1927, 6, (1), pp. 165–189
[13] Landsberg, J.M.: ‘Tensors: geometry and applications’, in David, Cox (Ed.):
Table 1 Summary of the tensor decompositions and ‘Graduate studies in mathematics’ (American Mathematical Society,
algorithms used in the applications Providence, Rhode Island, USA 2012), vol. 128
[14] Hackbusch, W.: ‘Tensor spaces and numerical tensor Calculus’, in Randolph
Application Tensor decompositions Algorithms E., Bank (Ed.): ‘Series in computational mathematics’ (Springer, Berlin,
DOA Canonical polyadic alternating least squares Heidelberg, 2012)
decomposition [15] Kolda, T.G., Bader, B.W.: ‘Tensor decompositions and applications’, SIAM
Rev., 2009, 51, (3), pp. 455–500
MHR generalised Vandermonde rectified ALS (RecALS), [16] Kruskal, J.B.: ‘Three-way arrays: rank and uniqueness of trilinear
CPD Vandermonde tensor-train decompositions’, Linear Algebra Appl., 1977, 18, pp. 95–138
RecALS (VTT-RecALS) [17] Sidiropoulos, N.D., Bro, R.: ‘On the uniqueness of multilinear decomposition
of N-way arrays’, J. Chemo, 2000, 14, pp. 229–239
MIMO CPD, Tucker- N1, N , Kronecker factorisation [18] Stegeman, A., Sidiropoulos, N.: ‘On Kruskal's uniqueness condition for the
nested tucker train (NTT) (KF), Khatri–Rao CP decomposition’, Linear Algebra Appl., 2007, 420, (2–3), pp. 540–552
factorisation (KRF) [19] Comon, P., Mourrain, B., Lim, L.H., et al.: ‘Genericity and rank deficiency of
high order symmetric tensors’. 2006 IEEE Int. Conf. on Acoustics Speech and
Signal Processing Proc., Toulouse, 2006
[20] Comon, P., Berge, J.M.F.T., De Lathauwer, L., et al.: ‘Generic and typical
respectively, before transmission. In this case, the noiseless ranks of multi-way arrays’, Linear. Algebra. Appl., 2009, 430, (11–12), pp.
received signal tensor at the destination node reduces to 2997–3007
[21] Chiantini, L., Ottaviani, G., Vannieuwenhoven, N.: ‘An algorithm for generic
and low-rank specific identifiability of complex tensors’, SIAM J. matrix
QT M Anal. Appl., 2014, 35, (4), pp. 1265–1287
xk, j, p, n = ∑ ∑ hkRDq cqR jhqSRmcmS psn r,
(
, T
) ( )
T,
(
R,
) ( )
, , (114) [22] Qi, Y., Comon, P., Lim, L.H.: ‘Uniqueness of nonnegative tensor
qT = 1 m = 1 approximations’, IEEE Trans. Inf. Theory, 2016, 62, (4), pp. 2170–2183,
arXiv:1410.8129
[23] Comon, P., Luciani, X., De Almeida, A.L.F.: ‘Tensor decompositions,
which corresponds to a nested CP model. This can be concluded by alternating least squares and other tales’, J. Chemometrics, 2009, 23, (7–8),
comparing (114) with (50). pp. 393–405
Assuming that the coding tensors C(S) and C(R) are known at the [24] Lim, L.H., Comon, P.: ‘Blind multilinear identification’, IEEE Trans. Inf.
Theory, 2014, 60, (2), pp. 1260–1280
destination node (i.e. by the receiver), semi-blind receivers for [25] Sahnoun, S., Comon, P.: ‘Joint source estimation and localization’, IEEE
jointly estimating the symbol matrix S and the individual Trans. Signal. Proc., 2015, 63, (10), pp. 2485–2495. hal-01005352
[26] Lim, L.H., Comon, P.: ‘Nonnegative approximations of nonnegative tensors’,
channels H(SR), H(RD) are proposed in [31]. We refer the interested J. Chemometrics, 2009, 23, pp. 432–441
reader to this reference for further details on the receiver [27] Qi, Y., Comon, P., Lim, L.H.: ‘Semialgebraic geometry of nonnegative tensor
algorithms. rank’, SIAM J. Matrix Anal. Appl., 2016, 37, (4), pp. 1556–1580.
hal-01763832
We summarised in Table 1 the different tensor decompositions [28] Oseledets, I.V.: ‘Tensor-train decomposition’, SIAM J Sci. Comput., 2011, 33,
and algorithms used in the three applications presented in this (5), pp. 2295–2317
paper. [29] Oseledets, I.V., Tyrtyshnikov, E.E.: ‘Breaking the curse of dimensionality, or
how to use SVD in many dimensions’, SIAM J. Sci. Comput., 2009, 31, pp.
3744–3759
7 Conclusion [30] Zniyed, Y., Boyer, R., de Almeida, A.L.F., et al.: ‘A TT-based hierarchical
framework for decomposing high-order tensors’, SIAM J. Sci. Comput., 2020,
Collecting data from multisensor signal processing can be naturally 42, (2), pp. A822–A848
interpreted in the multilinear algebra context. In this work, three [31] Favier, G., Fernandes, C., de Almeida, A.: ‘Nested Tucker tensor
important array signal processing-based applications are exposed in decomposition with application to MIMO relay systems using tensor space-
time coding (TSTC)’, Elsevier Signal Process., 2016, 128, (4), pp. 318–331
which the tensor-based framework has proven to be useful. [32] Rocha, D., Favier, G., Fernandes, C.: ‘Closed-form receiver for multi-hop
Namely, the problems of interest are (i) array signal processing for MIMO relay systems with tensor space-time coding’, J. Commun. Inf. Syst.,
localisation, (ii) MHR and (iii) MIMO processing for wireless 2019, 34, (1), pp. 50–54
communications. In this overview paper, we present systematically [33] Favier, G., da Costa, M.N., de Almeida, A., et al.: ‘Tensor space-time (TST)
coding for (MIMO) wireless communication systems’, Elsevier Signal
in a synthetic way the necessary mathematical material to ensure a Process., 2012, 92, (4), pp. 1079–1092
good understanding of the problems of interest. [34] de Almeida, A., Favier, G.: ‘Double Khatri-Rao space-time-frequency coding
using semi-blind PARAFAC based receiver’, IEEE Signal Process. Lett.,
2013, 20, (5), pp. 471–474
8 References [35] Ximenes, L., Favier, G., de Almeida, A.: ‘Closed-form semi-blind receiver for
[1] Harshman, R.A.: ‘Foundations of the parafac procedure: models and MIMO relay systems using double Khatri-Rao space-time coding’, IEEE
conditions for an explanatory multimodal factor analysis’, UCLA Work. Pap. Signal Process. Lett., 2016, 23, (3), pp. 316–320
Phonetics, 1970, 16, pp. 1–84 [36] Zniyed, Y., Boyer, R., de Almeida, A.L.F., et al.: ‘High-order tensor
[2] Carroll, J.D., Chang, J.J.: ‘Analysis of individual differences in estimation via trains of coupled third-order CP and Tucker decompositions’,
multidimensional scaling via N-way generalization of Eckart-Young Linear Algebr. Appl., 2020, 588, pp. 304–337
decomposition’, Psychometrika, 1970, 35, (3), pp. 283–319 [37] Zniyed, Y., Boyer, R., de Almeida, A.L.F., et al.: ‘High-order CPD estimation
[3] Acar, E., Yener, B.: ‘Unsupervised multiway data analysis: a literature with dimensionality reduction using a tensor train model’. Proc. of the 26th
survey’, IEEE Trans. Knowl. Data Eng., 2008, 21, (1), pp. 6–20 European Signal Processing Conf., Rome, Italy, 2018
[4] Sidiropoulos, N., De Lathauwer, L., Fu, X., et al.: ‘Tensor decomposition for [38] Van.Loan, C.F., Pitsianis, N.: ‘Approximation with Kronecker products’,
signal processing andmachine learning’, IEEE Trans. Signal Process., 2017, Moonen, M.S., Golub, G.H., De.Moor, B.L.R. (Eds).: ‘Linear algebra for
65, pp. 3551–3582 large scale and real-time applications’. (Springer Netherlands, Dordrecht,
[5] Sidiropoulos, N.D., Bro, R., Giannakis, G.B.: ‘Parallel factor analysis in 1993), pp. 293–314
sensor array processing’, IEEE Trans. Signal Process., 2000, 48, (8), pp. [39] Wu, K.K., Yam, Y., Meng, H., et al.: ‘Kronecker product approximation with
2377–2388 multiple factor matrices via the tensor product algorithm’. Proc. IEEE Int.
[6] Jiang, T., Sidiropoulos, N.D., Berge, J.M.F.T.: ‘Almost-sure identifiability of Conf. on Systems, Man, and Cybernetics (SMC 2016), Budapest, Hungary,
multidimensional harmonic retrieval’, IEEE Trans. Signal Process., 2001, 49, 2016, pp. 004277–004282
(9), pp. 1849–1859 [40] Zhang, T., Golub, G.H.: ‘Rank-one approximation to high order tensors’,
[7] Sidiropoulos, N.D., Bro, R., Giannakis, G.B.: ‘Blind PARAFAC receivers for SIAM J. Matrix Anal. Appl., 2001, 23, (2), pp. 534–550
DSCDMA systems’, IEEE Trans. Signal Process., 2000, 48, (3), pp. 810–823 [41] Kofidis, E., Regalia, P.A.: ‘On the best rank-1 approximation of higher-order
supersymmetric tensors’, SIAM J. Matrix Anal. Appl., 2002, 23, (3), pp. 863–
884

708 IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709
© The Institution of Engineering and Technology 2021
17519683, 2020, 10, Downloaded from https://ietresearch.onlinelibrary.wiley.com/doi/10.1049/iet-spr.2020.0373 by Indian Institute Of Technology Guwahati, Wiley Online Library on [27/02/2024]. See the Terms and Conditions (https://onlinelibrary.wiley.com/terms-and-conditions) on Wiley Online Library for rules of use; OA articles are governed by the applicable Creative Commons License
[42] da Silva, A.P., Comon, P., de Almeida, A.L.F.: ‘A finite algorithm to compute [71] Kroonenberg, P.M., de Leeuw, J.: ‘Principal component analysis of three-
rank-1 tensor approximations’, IEEE Signal Process. Lett., 2016, 23, (7), pp. mode data by means of alternating least squares algorithms’, Psychometrika,
959–963 1980, 45, pp. 69–97
[43] Kibangou, A.Y., Favier, G.: ‘Non-iterative solution for PARAFAC with a [72] Clark, M.P., Scharf, L.L.: ‘Two-dimensional modal analysis based on
Toeplitz matrix factor’. Proc. EUSIPCO, Glasgow Scotland, 2009 maximum likelihood’, IEEE Trans. Signal Process., 1994, 42, pp. 1443–1452
[44] Roy, R., Paulraj, A., Kailath, T.: ‘ESPRIT–a subspace rotation approach to [73] Boyer, R.: ‘Deterministic asymptotic Cramér-Rao bound for the
estimation of parameters of cisoids in noise’, IEEE Trans. Acoust. Speech multidimensional harmonic model’, Signal Process., 2008, 88, pp. 2869–2877
Signal Process., 1986, 34, (5), pp. 1340–1342 [74] Sahnoun, S., Usevich, K., Comon, P.: ‘Multidimensional ESPRIT for damped
[45] Roy, R., Kailath, T.: ‘ESPRIT-estimation of signal parameters via rotational and undamped signals: algorithm, computations and perturbation analysis’,
invariance techniques’, IEEE Trans. Acoust. Speech Signal Process., 1989, IEEE Trans. Signal. Proc., 2017, 65, (22), pp. 5897–5910. hal-01360438
37, (7), pp. 984–995 [75] Liu, J., Liu, X.: ‘An eigenvector-based approach for multidimensional
[46] Bienvenu, G., Kopp, L.: ‘Principe de la goniométrie passive adaptative’. 1979 frequency estimation with improved identifiability’, IEEE Trans. Signal
7ème Colloque sur le traitement du signal et des images (GRETSI). (Groupe Process., 2006, 54, pp. 4543–4556
d'Etudes du Traitement du Signal et des Images, 1979 [76] Sorensen, M., De Lathauwer, L.: ‘Blind signal separation via tensor
[47] Schmidt, R.O.: ‘A signal subspace approach to multiple emitter location and decomposition with vandermonde factor: canonical polyadic decomposition’,
spectral estimation’, Stanford Univ, Stanford, CA, 1981 IEEE Trans. Signal Process., 2013, 61, pp. 5507–5519
[48] Miron, S., Song, Y., Brie, D., et al.: ‘Multilinear direction finding for sensor- [77] Bro, R., Sidiropoulos, N.D., Giannakis, G.B.: ‘A fast least squares algorithm
array with multiple scales of invariance’, IEEE Trans. Aerosp. Electron. Syst., for separating trilinear mixtures’. Int. Workshop on Independent Component
2015, 51, (3), pp. 2057–2070 Analysis and Blind Separationg, Aussois, France, Jan. 11–15, 1999
[49] Guo, X., Miron, S., Brie, D., et al.: ‘A CANDECOMP/PARAFAC perspective [78] Papy, J.M., De Lathauwer, L., Huffel, S.V.: ‘Exponential data fitting using
on uniqueness of DOA estimation using a vector sensor array’, IEEE Trans. multilinear algebra: the single-channel and multi-channel case’, Wiley Online
Signal Process., 2011, 59, (7), pp. 3475–3481 Library, 2005, 12, pp. 809–826
[50] Sørensen, M., De Lathauwer, L.: ‘Multiple invariance ESPRIT for [79] Goulart, J.H., Boizard, M., Boyer, R., et al.: ‘Tensor CP decomposition with
nonuniform linear arrays: a coupled canonical polyadic decomposition structured factor matrices: algorithms and performance’, IEEE. J. Sel. Top.
approach’, IEEE Trans. Signal Process., 2016, 64, (14), pp. 3693–3704 Signal. Process., 2016, 10, (4), pp. 757–769
[51] Liang, J., Yang, S., Zhang, J., et al.: ‘4D near-field source localization using [80] Markovsky, I.: ‘Low rank approximation: algorithms, implementation,
cumulant’, EURASIP J. Adv. Signal Process., 2007, 2007, pp. 1–10 applications’ (Springer Science & Business Media, 2011)
[52] Haardt, M., Roemer, F., Del.Galdo, G.: ‘Higher-order SVD-based subspace [81] Boyle, J.P., Dykstra, R.L.: ‘A method for finding projections onto the
estimation to improve the parameter estimation accuracy in multidimensional intersection of convex sets in hilbert spaces’. Advances in order restricted
harmonic retrieval problems’, IEEE Trans. Signal Process., 2008, 56, (7), pp. statistical inference, Springer, New York, NY, USA, 1986, pp. 28–47
3198–3213 [82] Rife, D.C., Boorstyn, R.R.: ‘Single tone parameter estimation from discrete-
[53] Boizard, M., Ginolhac, G., Pascal, F., et al.: ‘Numerical performance of a time observations’, IEEE Trans. Inf. Theory, 1974, 20, (5), pp. 591–598
tensorMUSIC algorithm based on HOSVD for a mixture of polarized [83] Boyer, R., Comon, P.: ‘Rectified ALS algorithm for multidimensional
sources’. 2013 21st European Signal Processing Conf. (EUSIPCO), harmonic retrieval’. Sensor Array and Multichannel Signal Processing
Marrakech Morocco, 2013 Workshop (SAM), Rio de Janeiro, Brazil, 2016
[54] Sidiropoulos, N.D.: ‘Generalizing carathéodory's uniqueness of harmonic [84] Sørensen, M., De Lathauwer, L.: ‘New uniqueness conditions for the
parameterization to n dimensions’, IEEE Trans. Inf. Theory, 2001, 47, pp. canonical polyadic decomposition of third-order tensors’, SIAM J.Matrix
1687–1690 Anal. Appl., 2015, 36, pp. 1381–1403
[55] Li, Y., Razavilar, J., Liu, K.J.R.: ‘A high-resolution technique for [85] Zheng, L., Tse, D.N.C.: ‘Diversity and multiplexing: A fundamental tradeoff
multidimensional NMR spectroscopie’, IEEE Trans. Biomed. Eng., 1998, 45, in multiple-antenna channels’, IEEE Trans. Signal Process., 2003, 49, (5), pp.
pp. 78–86 1073–1096
[56] Nion, D., Sidiropoulos, N.D.: ‘Tensor algebra and multidimensional harmonic [86] Favier, G., de Almeida, A.: ‘Tensor space-time-frequency coding with
retrieval in signal processing for MIMO radars’, IEEE Trans. Signal Process., semiblind receivers for MIMO wireless communication systems’, IEEE
2010, 58, pp. 5693–57057 Trans. Signal Process., 2014, 62, (22), pp. 5987–6002
[57] Zoltowski, M.D., Haardt, M., Mathews, C.P.: ‘Closed-form 2D angle [87] Liu, K., da Costa, J.P.C., So, H., et al.: ‘Semi-blind receivers for joint symbol
estimation with rectangular arrays in element space or beamspace via unitary and channel estimation in space-time-frequency MIMO-OFDM systems’,
ESPRIT’, IEEE Trans. Signal Process., 1996, 44, pp. 316–328 IEEE Trans. Signal Process., 2013, 61, (21), pp. 5444–5457
[58] Acar, E., Bro, R., Smilde, A.K.: ‘Data fusion in metabolomics using coupled [88] de Almeida, A., Favier, G., Mota, J.C.: ‘A constrained factor decomposition
matrix and tensor factorizations’, Proc. IEEE, 2015, 103, pp. 1602–1620 with application to MIMO antenna systems’, IEEE Trans. Signal Process.,
[59] Farias, R.C., Cohen, J.E., Comon, P.: ‘Exploring multimodal data fusion 2008, 56, (6), pp. 2429–2442
through joint decompositions with flexible couplings’, IEEE Trans. Signal [89] de Almeida, A., Favier, G., Mota, J.C.: ‘PARAFAC-based unified tensor
Process., 2016, 64, pp. 4830–4844 modeling for wireless communication systems with application to blind
[60] Lahat, D., Adali, T., Jutten, C.: ‘Multimodal data fusion: an overview of multiuser equalization’, Signal Process., 2007, 87, (2), pp. 337–351
methods, challenges and prospects’, Proc. IEEE, 2015, 103, pp. 1449–1477 [90] da Costa, M.N., Favier, G., Romano, J.M.T.: ‘Tensor modelling of MIMO
[61] Cichocki, A.: ‘Era of big data processing: A new approach via tensor communication systems with performance analysis and kronecker receivers’,
networks and tensor decompositions’. Int. Workshop on Smart Info-Media Elsevier Signal Process., 2018, 145, (4), pp. 304–316
Systems in Asia, Nagoya, Japan, 2013 [91] Rong, Y., Khandaker, M.R.A., Xiang, Y.: ‘Channel estimation of dual-hop
[62] Phan, A., Cichocki, A., Uschmajew, A., et al.: ‘Tensor networks for latent MIMO relay system via parallel factor analysis’, IEEE Trans. Wirel.
variable analysis. Part i: algorithms for tensor train decomposition’, IEEE Commun., 2012, 11, (6), pp. 2224–2233
Trans. Neural Netw. Learn. Syst., 2020, 31, (11), pp. 4622–4636 [92] Cavalcante, I.V., de Almeida, A.L.F., Haardt, M.: ‘Tensor-based approach to
[63] Vervliet, N., Debals, O., Sorber, L., et al.: ‘Breaking the curse of channel estimation in amplify-and-forward MIMO relaying systems’. Proc.
dimensionality using decompositions of incomplete tensors: tensor-based IEEE 8th Sensor Array and Multichannel Signal Processing Workshop (SAM
scientific computing in big data analysis’, SIAM J. Sci. Comput., 2014, 31, pp. 2014), Coruna, Spain, 2014, pp. 445–448
71–79 [93] Roemer, F., Haardt, M.: ‘Tensor-based channel estimation and iterative
[64] Hackbusch, W., Kuhn, S.: ‘A new scheme for the tensor representation’, J. refinements for two-way relaying with multiple antennas and spatial reuse’,
Fourier Anal. Appl., 2009, 15, pp. 706–722 IEEE Trans. Signal Process., 2010, 58, (11), pp. 5720–5735
[65] Kressner, D., Steinlechner, M., Vandereycken, B.: ‘Low-rank tensor [94] Fernandes, C.A.R., de Almeida, A.L.F., da Costa, D.B.: ‘Unified tensor
completion by Riemannian optimization’, BIT Numerical Math., 2014, 54, pp. modeling for blind receivers in multiuser uplink cooperative systems’, IEEE
447–468 Signal Process. Lett., 2012, 19, (5), pp. 247–250
[66] Bousse, M., Debals, O., De Lathauwer, L.: ‘A tensor-based method for large- [95] de Almeida, A.L.F., Fernandes, C.A.R., da Costa, D.B.: ‘Multiuser detection
scale blind source separation using segmentation’, IEEE Trans. Signal for uplink DS-CDMA amplify-and-forward relaying systems’, IEEE Signal
Process., 2016, 65, pp. 346–358 Process. Lett., 2013, 20, (7), pp. 697–700
[67] Roemer, F., Haardt, M.: ‘A closed-form solution for parallel factor [96] Ximenes, L.R., Favier, G., de Almeida, A.L.F.: ‘Semi-blind receivers for non-
(PARAFAC) analysis’. IEEE Int. Conf. on Acoustics, Speech and Signal regenerative cooperative MIMO communications based on nested PARAFAC
Processing, Las Vegas, USA, 2008 modeling’, IEEE Trans. Signal Process., 2015, 63, (18), pp. 4985–4998
[68] Bro, R.: ‘PARAFAC: tutorial and applications’, Chemometr. Intell. Lab. Syst., [97] Sokal, B., de Almeida, A.L.F., Haardt, M.: ‘Semi-blind receivers for MIMO
1997, 38, pp. 149–171 multi-relaying systems via rank-one tensor approximations’, Signal Process.,
[69] Li, N., Kindermann, S., Navasca, C.: ‘Some convergence results on the 2020, 166, p. 107254
regularized alternating least-squares method for tensor decomposition’, [98] Rocha, D.S., Fernandes, C., Favier, G.: ‘MIMO multi-relay systems with
Linear Algebr. Appl., 2013, 438, pp. 796–812 tensor space-time coding based on coupled nested tucker decomposition’,
[70] Cichocki, A. N., Lee, I.V.O., Phan, A.H., et al.: ‘Low-rank tensor networks Digit. Signal Process., 2019, 89, (3), pp. 170–185
for dimensionality reduction and large-scale optimization problems: [99] Sidiropoulos, N.D., Budampati, R.: ‘Khatri-Rao space-time codes’, IEEE
perspectives and challenges’, Foundations and Trends in Mach. Learn., 2016, Trans. Signal Process., 2002, 50, (10), pp. 2396–2407
9, pp. 249–420

IET Signal Process., 2020, Vol. 14 Iss. 10, pp. 693-709 709
© The Institution of Engineering and Technology 2021

You might also like