Tensors for Data Processing. Theory, Methods, and Applications Yipeng Liuinstant download
Tensors for Data Processing. Theory, Methods, and Applications Yipeng Liuinstant download
or textbooks at https://ebookmass.com
_____ Follow the link below to get your download now _____
https://ebookmass.com/product/tensors-for-data-processing-
theory-methods-and-applications-yipeng-liu/
https://ebookmass.com/product/tensors-for-data-processing-1st-edition-
yipeng-liu/
https://ebookmass.com/product/spatial-analysis-using-big-data-methods-
and-urban-applications-yamagata/
https://ebookmass.com/product/magnetic-communications-theory-and-
techniques-liu/
https://ebookmass.com/product/processing-for-android-create-mobile-
sensor-aware-and-xr-applications-using-processing-2nd-edition-colubri/
https://ebookmass.com/product/big-data-in-astronomy-scientific-data-
processing-for-advanced-radio-telescopes-1st-edition-linghe-kong-
editor/
https://ebookmass.com/product/oil-palm-biomass-for-composite-panels-
fundamentals-processing-and-applications-s-m-sapuan/
Edited by
Yipeng Liu
School of Information and Communication Engineering
University of Electronic Science and Technology
of China (UESTC)
Chengdu, China
Academic Press is an imprint of Elsevier
125 London Wall, London EC2Y 5AS, United Kingdom
525 B Street, Suite 1650, San Diego, CA 92101, United States
50 Hampshire Street, 5th Floor, Cambridge, MA 02139, United States
The Boulevard, Langford Lane, Kidlington, Oxford OX5 1GB, United Kingdom
Copyright © 2022 Elsevier Inc. All rights reserved.
ISBN: 978-0-12-824447-0
v
vi Contents
9.4.4
Compressing the convolutional layer via TT/TR
decompositions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
9.4.5 Compressing neural networks via transform-based
decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 330
9.5 Experiments and future directions . . . . . . . . . . . . . . . . . . . . . 333
9.5.1 Performance evaluations using the MNIST dataset . . . 333
9.5.2 Performance evaluations using the CIFAR10 dataset . 336
9.5.3 Future research directions . . . . . . . . . . . . . . . . . . . . . 337
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338
CHAPTER 10 Coupled tensor decompositions for data fusion . . . . 341
Christos Chatzichristos, Simon Van Eyndhoven,
Eleftherios Kofidis, and Sabine Van Huffel
10.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 341
10.2 What is data fusion? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342
10.2.1 Context and definition . . . . . . . . . . . . . . . . . . . . . . . . 342
10.2.2 Challenges of data fusion . . . . . . . . . . . . . . . . . . . . . . 343
10.2.3 Types of fusion and data fusion strategies . . . . . . . . . . 347
10.3 Decompositions in data fusion . . . . . . . . . . . . . . . . . . . . . . . . 348
10.3.1 Matrix decompositions and statistical models . . . . . . . 350
10.3.2 Tensor decompositions . . . . . . . . . . . . . . . . . . . . . . . 351
10.3.3 Coupled tensor decompositions . . . . . . . . . . . . . . . . . 352
10.4 Applications of tensor-based data fusion . . . . . . . . . . . . . . . . 355
10.4.1 Biomedical applications . . . . . . . . . . . . . . . . . . . . . . . 355
10.4.2 Image fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
10.5 Fusion of EEG and fMRI: a case study . . . . . . . . . . . . . . . . . . 358
10.6 Data fusion demos . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
10.6.1 SDF demo – approximate coupling . . . . . . . . . . . . . . 361
10.7 Conclusion and prospects . . . . . . . . . . . . . . . . . . . . . . . . . . . . 363
Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
CHAPTER 11 Tensor methods for low-level vision . . . . . . . . . . . . . . 371
Tatsuya Yokota, Cesar F. Caiafa, and Qibin Zhao
11.1 Low-level vision and signal reconstruction . . . . . . . . . . . . . . . 371
11.1.1 Observation models . . . . . . . . . . . . . . . . . . . . . . . . . . 372
11.1.2 Inverse problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374
11.2 Methods using raw tensor structure . . . . . . . . . . . . . . . . . . . . 378
11.2.1 Penalty-based tensor reconstruction . . . . . . . . . . . . . . 379
11.2.2 Tensor decomposition and reconstruction . . . . . . . . . . 393
11.3 Methods using tensorization . . . . . . . . . . . . . . . . . . . . . . . . . . 409
11.3.1 Higher-order tensorization . . . . . . . . . . . . . . . . . . . . . 411
11.3.2 Delay embedding/Hankelization . . . . . . . . . . . . . . . . 413
11.4 Examples of low-level vision applications . . . . . . . . . . . . . . . 415
Contents xi
Kim Batselier
Delft Center for Systems and Control, Delft University of Technology, Delft,
The Netherlands
Yingyue Bi
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
Jérémie Boulanger
CRIStAL, Université de Lille, Villeneuve d’Ascq, France
Rémy Boyer
CRIStAL, Université de Lille, Villeneuve d’Ascq, France
Cesar F. Caiafa
Instituto Argentino de Radioastronomía – CCT La Plata, CONICET / CIC-PBA /
UNLP, Villa Elisa, Argentina
RIKEN Center for Advanced Intelligence Project, Tokyo, Japan
Jocelyn Chanussot
LJK, CNRS, Grenoble INP, Inria, Université Grenoble, Alpes, Grenoble, France
Christos Chatzichristos
KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for
Dynamical Systems, Signal Processing and Data Analytics, Leuven, Belgium
Cong Chen
Department of Electrical and Electronic Engineering, The University of Hong
Kong, Pokfulam Road, Hong Kong
Nadav Cohen
School of Computer Science, Hebrew University of Jerusalem, Jerusalem, Israel
Xudong Cui
School of Mathematics, Tianjin University, Tianjin, China
André L.F. de Almeida
Department of Teleinformatics Engineering, Federal University of Fortaleza,
Fortaleza, Brazil
Aybüke Erol
Circuits and Systems, Department of Microelectronics, Delft University of
Technology, Delft, The Netherlands
xiii
xiv List of contributors
Yiming Fang
Department of Computer Science, Columbia University, New York, NY,
United States
Gérard Favier
Laboratoire I3S, Université Côte d’Azur, CNRS, Sophia Antipolis, France
Borbála Hunyadi
Circuits and Systems, Department of Microelectronics, Delft University of
Technology, Delft, The Netherlands
Pratik Jawanpuria
Microsoft, Hyderabad, India
Tai-Xiang Jiang
School of Economic Information Engineering, Southwestern University of
Finance and Economics, Chengdu, Sichuan, China
Paris A. Karakasis
School of Electrical and Computer Engineering, Technical University of Crete,
Chania, Greece
Ouafae Karmouda
CRIStAL, Université de Lille, Villeneuve d’Ascq, France
Hiroyuki Kasai
Waseda University, Tokyo, Japan
Eleftherios Kofidis
Dept. of Statistics and Insurance Science, University of Piraeus, Piraeus, Greece
Christos Kolomvakis
School of Electrical and Computer Engineering, Technical University of Crete,
Chania, Greece
Yoav Levine
School of Computer Science, Hebrew University of Jerusalem, Jerusalem, Israel
Zechu Li
Department of Computer Science, Columbia University, New York, NY,
United States
Athanasios P. Liavas
School of Electrical and Computer Engineering, Technical University of Crete,
Chania, Greece
List of contributors xv
Zhouchen Lin
Key Lab. of Machine Perception, School of EECS, Peking University, Beijing,
China
Xiao-Yang Liu
Department of Computer Science and Engineering, Shanghai Jiao Tong
University, Shanghai, China
Department of Electrical Engineering, Columbia University, New York, NY,
United States
Yipeng Liu
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
Zhen Long
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
George Lourakis
Neurocom, S.A, Athens, Greece
Canyi Lu
Carnegie Mellon University, Pittsburgh, PA, United States
Liangfu Lu
School of Mathematics, Tianjin University, Tianjin, China
Yingcong Lu
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
George Lykoudis
Neurocom, S.A, Athens, Greece
Bamdev Mishra
Microsoft, Hyderabad, India
Michael K. Ng
Department of Mathematics, The University of Hong Kong, Pokfulam,
Hong Kong
Ioannis Marios Papagiannakos
School of Electrical and Computer Engineering, Technical University of Crete,
Chania, Greece
Bo Ren
Key Laboratory of Intelligent Perception and Image Understanding of Ministry of
Education of China, Xidian University, Xi’an, China
xvi List of contributors
Or Sharir
School of Computer Science, Hebrew University of Jerusalem, Jerusalem, Israel
Amnon Shashua
School of Computer Science, Hebrew University of Jerusalem, Jerusalem, Israel
Ioanna Siaminou
School of Electrical and Computer Engineering, Technical University of Crete,
Chania, Greece
Christos Tsalidis
Neurocom, S.A, Athens, Greece
Simon Van Eyndhoven
KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for
Dynamical Systems, Signal Processing and Data Analytics, Leuven, Belgium
icometrix, Leuven, Belgium
Sabine Van Huffel
KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for
Dynamical Systems, Signal Processing and Data Analytics, Leuven, Belgium
Anwar Walid
Nokia Bell Labs, Murray Hill, NJ, United States
Fei Wen
Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai,
China
Noam Wies
School of Computer Science, Hebrew University of Jerusalem, Jerusalem, Israel
Ngai Wong
Department of Electrical and Electronic Engineering, The University of Hong
Kong, Pokfulam Road, Hong Kong
Zebin Wu
School of Computer Science and Engineering, Nanjing University of Science
and Technology, Nanjing, China
Yang Xu
School of Computer Science and Engineering, Nanjing University of Science
and Technology, Nanjing, China
Liuqing Yang
Department of Computer Science, Columbia University, New York, NY,
United States
List of contributors xvii
Fei Ye
School of Computer Science and Engineering, Nanjing University of Science
and Technology, Nanjing, China
Tatsuya Yokota
Nagoya Institute of Technology, Aichi, Japan
RIKEN Center for Advanced Intelligence Project, Tokyo, Japan
Zhonghao Zhang
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
Qibin Zhao
RIKEN Center for Advanced Intelligence Project, Tokyo, Japan
Guangdong University of Technology, Guangzhou, China
Xi-Le Zhao
School of Mathematical Sciences/Research Center for Image and Vision
Computing, University of Electronic Science and Technology of China, Chengdu,
Sichuan, China
Pan Zhou
SEA AI Lab, Singapore, Singapore
Ce Zhu
School of Information and Communication Engineering, University of Electronic
Science and Technology of China (UESTC), Chengdu, China
Yassine Zniyed
Université de Toulon, Aix-Marseille Université, CNRS, LIS, Toulon, France
This page intentionally left blank
Preface
This book provides an overview of tensors for data processing, covering computing
theories, processing methods, and engineering applications. The tensor extensions
of a series of classical multidimensional data processing techniques are discussed
in this book. Many thanks go to all the contributors. Students can read this book to
get an overall understanding, researchers can update their knowledge on the recent
research advances in the field, and engineers can refer to implementations on various
applications.
The first chapter is an introduction to tensor decomposition. In the following, the
book provides variants of tensor decompositions with their efficient and effective so-
lutions, including some parallel algorithms, Riemannian algorithms, and generalized
thresholding algorithms. Some tensor-based machine learning methods are summa-
rized in detail, including tensor completion, tensor principal component analysis,
support tensor machine, tensor-based kernel learning, tensor-based deep learning, etc.
To demonstrate that tensors can effectively and systematically enhance performance
in practical engineering problems, this book gives implemental details of many ap-
plications, such as signal recovery, recommender systems, climate forecasting, image
clustering, image classification, network compression, data fusion, image enhance-
ment, neuroimaging, and remote sensing.
I sincerely hope this book can serve to introduce tensors to more data scientists
and engineers. As a natural representation of multidimensional data, tensors can be
used to substantially avoid the information loss in matrix representations of multiway
data, and tensor operators can model more connections than their matrix counterparts.
The related advances in applied mathematics allow us to move from matrices to ten-
sors for data processing. This book is promising to motivate novel tensor theories and
new data processing methods, and to stimulate the development of a wide range of
practical applications.
Yipeng Liu
Chengdu, China
Aug. 10, 2021
xix
This page intentionally left blank
CHAPTER
Tensor decompositions:
computations,
applications, and
challenges
1
Yingyue Bi, Yingcong Lu, Zhen Long, Ce Zhu, and Yipeng Liu
School of Information and Communication Engineering, University of Electronic Science and
Technology of China (UESTC), Chengdu, China
CONTENTS
1.1 Introduction..................................................................................... 1
1.1.1 What is a tensor? ............................................................... 1
1.1.2 Why do we need tensors? ..................................................... 2
1.2 Tensor operations ............................................................................. 3
1.2.1 Tensor notations ................................................................ 3
1.2.2 Matrix operators ................................................................ 4
1.2.3 Tensor transformations ........................................................ 6
1.2.4 Tensor products................................................................. 7
1.2.5 Structural tensors .............................................................. 11
1.2.6 Summary ........................................................................ 13
1.3 Tensor decompositions ....................................................................... 13
1.3.1 Tucker decomposition ......................................................... 13
1.3.2 Canonical polyadic decomposition........................................... 14
1.3.3 Block term decomposition .................................................... 16
1.3.4 Tensor singular value decomposition ........................................ 18
1.3.5 Tensor network .................................................................. 19
1.4 Tensor processing techniques............................................................... 24
1.5 Challenges ...................................................................................... 25
References............................................................................................ 26
1.1 Introduction
1.1.1 What is a tensor?
The tensor can be seen as a higher-order generalization of vector and matrix, which
normally has three or more modes (ways) [1]. For example, a color image is a third-
order tensor. It has two spatial modes and one channel mode. Similarly, a color video
is a fourth-order tensor; its extra mode denotes time.
Tensors for Data Processing. https://doi.org/10.1016/B978-0-12-824447-0.00007-8
Copyright © 2022 Elsevier Inc. All rights reserved.
1
2 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.1
A third-order tensor A ∈ RI1 ×I2 ×I3 .
FIGURE 1.2
The illustration of mode-1 fibers A(:, i2 , i3 ), mode-2 fibers A(i1 , :, i3 ), and mode-3 fibers
A(i1 , i2 , :) with i1 = 1, · · · , I1 , i2 = 1, · · · , I2 and i3 = 1, · · · , I3 .
Other than the aforementioned notations, there is another way to denote tensors
and their operations [5]. Taking advantage of graphical representations, tensors can
be denoted by nodes and edges in a straightforward way. Graphical representations
for scalars, vectors, matrices, and tensors are shown in Fig. 1.4. The number next to
the edge represents the indices of the corresponding mode.
4 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.3
The illustration of horizontal slices A(i1 , :, :) i1 = 1, · · · , I1 , lateral slices A(:, i2 , :)
i2 = 1, · · · , I2 , and frontal slices A(:, :, i3 ) i3 = 1, · · · , I3 .
FIGURE 1.4
Graphical representations of scalar, vector, matrix and tensor.
Definition 1.2.2. (p -norm [6]) For matrix A ∈ RI ×J , its p -norm is defined as
1/p
I
J
p
Ap = ai,j . (1.1)
i=1 j =1
Definition 1.2.3. (Matrix nuclear norm [7]) The nuclear norm of matrix A is denoted
as A∗ = i σi (A), where σi (A) is the i-th largest singular value of A.
1.2 Tensor operations 5
Definition 1.2.4. (Hadamard product [8]) The Hadamard product for matrices A ∈
RM×N and B ∈ RM×N is defined as A B ∈ RM×N with
⎡ ⎤
a1,1 b1,1 a1,2 b1,2 ··· a1,N b1,N
⎢ a2,1 b2,1 a2,2 b2,2 ··· a2,N b2,N ⎥
⎢ ⎥
AB=⎢ .. .. .. .. ⎥. (1.2)
⎣ . . . . ⎦
aM,1 bM,1 aM,2 bM,2 ··· aM,N bM,N
Based on the Kronecker product, a lot of useful properties can be derived. Given
matrices A, B, C, D, we have
(A ⊗ B)(C ⊗ D) = AC ⊗ BD,
(A ⊗ B)† = A† ⊗ B† , (1.4)
(A ⊗ B) = A ⊗ B ,
T T T
Similar to the Kronecker product, the Khatri–Rao product also has some conve-
nient properties, such as
(A B)T = AT BT ,
A B C = (A B) C = A (B C),
(1.6)
(A B)T (A B) = AT A BT B,
†
(A B)† = AT A BT B (A B)T .
6 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.5
A graphical illustration of the tensor transpose on A ∈ RI1 ×I2 ×5 .
⎛ ⎞
1 2 9 2 8 7
A(2) = ⎝ 4 8 5 6 1 5 ⎠ , (1.9)
5 7 3 2 3 6
1 2 9 4 8 5 5 7 3
A(3) = . (1.10)
2 8 7 6 1 5 2 3 6
1.2 Tensor operations 7
FIGURE 1.6
A graphical illustration of tensor mode-n matricization for A ∈ RI1 ×···×IN .
Definition 1.2.9. (Tensor n-th canonical matricization [12]) For a fixed index n =
1, 2, · · · , N, the n-th canonical matricization of tensor A ∈ RI1 ×I2 ×···×IN can be de-
fined as
(A<n> )i1 i2 ···in , in+1 ···iN = ai1 ,i2 ,··· ,iN , (1.11)
where i1 i2 · · · in , in+1 · · · iN are multiindices and A<n> ∈ RI1 I2 ···In ×In+1 ···IN .
Take the multiindex i = i1 i2 · · · iN as an example, in = 1, 2, · · · , In , n = 1, · · · , N.
It can either be defined using the little-endian convention (reverse lexicographic or-
dering) [13]
I1
I2
IN
A, B = ··· ai1 ,i2 ,··· ,iN bi1 ,i2 ,··· ,iN . (1.14)
i1 =1 i2 =1 iN =1
Definition 1.2.11. (Tensor norm [1]) The norm of a tensor A ∈ RI1 ×I2 ×···×IN is the
square root of the summation over the square of all its elements, which can be ex-
pressed as
I1 I2
IN
A = ··· (ai1 ,i2 ,··· ,iN )2 . (1.15)
i1 =1 i2 =1 iN =1
8 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.7
A graphical illustration of the tensor inner product.
Definition 1.2.12. (Tensor mode-n product with a matrix [1]) The tensor mode-n
product of A ∈ RI1 ×I2 ×···×IN and matrix B ∈ RK×In is denoted as
or element-wisely,
IN
xi1 ,··· ,k,··· ,iN = ai1 ,··· ,in ,··· ,iN bk,in . (1.17)
in =1
FIGURE 1.8
A graphical illustration of the tensor mode-n product.
1.2 Tensor operations 9
1 2 3
For example, given tensor A (Eq. (1.7)) and matrix B = , the
4 5 6
mode-n product A ×1 B will yield a tensor X ∈ R2×3×2 , whose frontal slices are
32 35 28 39 23 26
X (:, :, 1) = , X (:, :, 2) = . (1.19)
68 86 73 90 59 59
Definition 1.2.13. (Tensor mode-n product with a vector [1]) The tensor mode-n
product of the tensor A ∈ RI1 ×I2 ×···×IN and vector b ∈ RIn is denoted as
with entries
IN
xi1 ,··· ,in−1 ,in+1 ,··· ,iN = ai1 ,··· ,in−1 ,in ,in+1 ,··· ,iN bin . (1.21)
in =1
T
For example, given tensor A in Eq. (1.7) and vector b = 1 2 3 , we have
⎡
⎤
24 20
A ×2 b = ⎣ 39 19 ⎦ . (1.22)
28 35
It can be clearly seen that the operation of multiplying a tensor by a matrix will
not change the number of ways of the tensor. However, if a tensor is multiplied by a
vector, the number of ways will decrease.
Definition 1.2.14. (t-product [11]) The t-product of A ∈ RI1 ×I2 ×I3 and C ∈ RI2 ×L×I3
is defined as
X = A ∗ C = fold circ(A)MatVec(C) , (1.23)
T
where X ∈ RI1 ×L×I3 , MatVec(C) = C(1)T C(2)T · · · C(I3 )T ∈ RI2 I3 ×L
represents the block matrix [11] of C, and
⎡ ⎤
A(1) A(I3 ) ··· A(2)
⎢ A(2) A(1) ... A(3) ⎥
⎢ ⎥
circ(A) = ⎢ .. .. .. .. ⎥ ∈ RI1 I3 ×I2 I3
⎣ . . . . ⎦
A(I3 ) A(I3 − 1) · · · A(1)
Definition 1.2.15. (Tensor contraction [5]) Given two tensors A ∈ RI1 ×I2 ×···×IM
and B ∈ RJ1 ×J2 ×···×JN , suppose they have L equal indices {K1 , K2 , · · · , KL } in
10 CHAPTER 1 TDs: computations, applications, and challenges
K1
KL
··· ai1 ,··· ,iM bj1 ,··· ,jN . (1.24)
k1 =1 kL =1
FIGURE 1.9
Graphical representation of contraction of two tensors, A ∈ RI1 ×···×IM and B ∈ RJ1 ×···×JN ,
where {K1 , K2 , · · · , KL } denotes the L equal indices in {I1 , I2 , · · · , IM } and
{J1 , J2 , · · · , JN }.
4
2
7
xi1 ,i4 ,j2 ,j4 = ai1 ,k1 ,k2 ,i4 ,k3 bk2 ,j2 ,k3 ,j4 ,k1 . (1.25)
k1 =1 k2 =1 k3 =1
FIGURE 1.10
The contraction of two tensors, A ∈ R3×4×2×6×7 and B ∈ R2×5×7×8×4 , where
K1 = I2 = J5 = 4, K2 = I3 = J1 = 2, K3 = I5 = J3 = 7, I1 = 3, I4 = 6, J2 = 5, and J4 = 8.
1.2 Tensor operations 11
FIGURE 1.11
A graphical representation of contraction over two tensors, A ∈ RI1 ×I2 ×···×IM and
B ∈ RJ1 ×J2 ×···×JN , where K1 = Im = Jn .
The contraction of tensors A ∈ RI1 ×I2 ×···×IM and B ∈ RJ1 ×J2 ×···×JN results in an
(M + N − 2)-th-order tensor X = A, B1 , whose entries can be calculated by
xi1 ,··· ,im−1 ,im+1 ,··· ,iM ,j1 ,··· ,jn−1 ,jn+1 ,··· ,jN
K1
= ai1 ,··· ,im−1 ,k1 ,im+1 ,··· ,iM bj1 ,··· ,jn−1 ,k1 ,jn+1 ,··· ,jN . (1.26)
k1 =1
we have
A = a(1) ◦ a(2) ◦ · · · ◦ a(N ) . (1.29)
FIGURE 1.12
A rank-1 tensor A = a(1) ◦ a(2) ◦ a(3) ∈ RI1 ×I2 ×I3 .
FIGURE 1.13
A third-order diagonal tensor A ∈ RI1 ×I2 ×I3 .
FIGURE 1.14
An f -diagonal tensor A ∈ RI1 ×I2 ×I3 .
1.2.6 Summary
In this section, we first briefly described some notations of tensor representations.
Then by giving basic operations of matrices, we discussed several common tensor
operations, including tensor transformations and tensor products. Concepts of struc-
tural tensors such as orthogonal tensor, diagonal tensor, and f -diagonal tensor are
also given. It is worth noting that we only focus on the most commonly used defini-
tions; for more information, please refer to [1], [5], and [6].
where U(n) ∈ RIn ×Rn (n = 1, · · · , N ) are semi-orthogonal factor matrices that satisfy
U(n)T U(n) = IRn and G ∈ RR1 ×R2 ×···×RN is the core tensor. Even though the core
tensor is usually dense, it is generally much smaller than X , i.e., Rn In .
We can also write Tucker decomposition in an element-wise style as
R1
R2
RN
(1) (2) (N )
xi1 ,i2 ,··· ,iN = ··· gr1 ,r2 ,··· ,rN ui1 ,r1 ui2 ,r2 · · · uiN ,rN , (1.31)
r1 r2 rN
14 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.15
An illustration of Tucker decomposition on a third-order tensor T . The core tensor is
G ∈ RR1 ×R2 ×R3 and factor matrices are A, B, C ∈ RIn ×Rn , n = 1, 2, 3.
Definition 1.3.2. (Tucker rank) The Tucker rank of a given tensor X ∈ RI1 ×···×IN
is defined as an N-tuple (R1 , · · · , RN ) comprised of n-rank Rn . The n-rank Rn =
rank(X(n) ), a.k.a. the multilinear rank, is the dimension of the vector space spanned
by the mode-n fibers. In other words, the n-rank is the column rank of X(n) .
T = G; A, B, C
= G ×1 R ×2 P ×3 Q; AR−1 , BP−1 , CQ−1 ,
R
X= r ◦ ur ◦ · · · ◦ ur
u(1) (2) (N )
= U(1) , U(2) , · · · , U(N ) , (1.32)
r=1
R
xi1 ,i2 ,··· ,iN = u(1) (2) (N )
i1 ,r ui2 ,r · · · uiN ,r . (1.33)
r=1
FIGURE 1.16
A demonstration of CP decomposition of a third-order tensor. Each dotted rectangle
represents a rank-1 tensor resulting from ar ◦ br ◦ cr , r = 1, · · · , R.
R
R
T = ar ◦ br ◦ cr = (αr ar ) ◦ (βr br ) ◦ (γr cr ), (1.36)
r=1 r=1
Definition 1.3.5. (Block term decomposition) For a tensor T ∈ RI1 ×I2 ×···×IN , its
BTD is denoted as
R
T = Sr ×1 W(1)
r ×2 Wr · · · ×N Wr ,
(2) (N )
(1.37)
r=1
(n)
where Wr ∈ RIn ×Mr represents the n-th factor in the r-th term and Sr ∈
(n)
(1) (2) (N)
RMr ×Mr ×···×Mr represents the corresponding core tensor.
Considering BTD on third-order tensors, we can give the definition of (L, M, N)-
decomposition.
R
T = Sr ×1 Ar ×2 Br ×3 Cr , (1.38)
r=1
FIGURE 1.17
The illustration of block term decomposition for a third-order tensor.
R
T = Ar BTr ◦ cr , (1.39)
r=1
However, Ar and Br may not have the same rank on some occasions. Therefore,
(Lr , Lr , 1)-decomposition was proposed, as shown in Fig. 1.18.
FIGURE 1.18
The illustration of (Lr , Lr , 1)-decomposition for a third-order tensor.
R
T = Sr ×1 Ar ×2 Br , (1.41)
r=1
FIGURE 1.19
The illustration of (L, M, ·)-decomposition for a third-order tensor.
Definition 1.3.11. (t-SVD) For tensor T ∈ RI1 ×I2 ×I3 , the t-SVD is defined as
T = U ∗ S ∗ V T, (1.42)
where U ∈ RI1 ×I1 ×I3 , V ∈ RI2 ×I2 ×I3 are orthogonal tensors and S ∈ RI1 ×I2 ×I3 is an
f -diagonal tensor.
FIGURE 1.20
The illustration of t-SVD for a third-order tensor.
Definition 1.3.12. (Tensor tubal rank [38]) The tubal rank of a third-order tensor T
is defined as the number of nonzero tubes in S.
It is well known that block-circulant matrices can be block-diagonalized by the
Fourier transform [39]. Mathematically, for T ∈ RI1 ×I2 ×I3 , we have
⎡ ⎤
T(1)
⎢ ⎥
(F ⊗ I)circ(T ) F∗ ⊗ I = bdiag(T ) = ⎢⎣
..
.
⎥,
⎦ (1.43)
T(I3 )
where circ(T ) is the block-circulant matrix (see Eq. (1.23)) of T , F ∈ RI3 ×I3 is a
normalized discrete Fourier transform (DFT) matrix, T is the fast Fourier transform
(FFT) of T along the third mode, and T(i3 ) ∈ RI1 ×I2 (i3 = 1, · · · , I3 ) represents its
i3 -th frontal slice.
In this way, instead of directly calculating circ(·), we are able to employ the FFT
to solve Eq. (1.42). Specifically, we apply SVD on every frontal slice of T to acquire
U(i3 ), S(i3 ), and V(i3 ), i3 = 1, · · · , I3 . Then by simply performing the inverse FFT
along the third mode on U, S, and V, the t-SVD of T is obtained. From this point
of view, t-SVD can be regarded as performing matrix SVD on each frontal slice of a
tensor in the frequency domain.
FIGURE 1.21
A possible binary tree T of a fourth-order tensor. Each node corresponds to a mode set of
the tensor A ∈ RI1 ×I2 ×I3 ×I4 .
where
ICq := Ic , IC̄q := Ic̄ .
c∈Cq c̄∈C̄q
For example, given tensor X and dimension indices Cq = C2 = {1, 2} in Fig. 1.21,
the matricization X[q] = X[2] is of the size of I1 I2 × I3 I4 .
Definition 1.3.14. (Hierarchical rank) The hierarchical rank of tensor X ∈ RI1 ×···×IN
is defined as
K = KCq |KCq = rank(X[q] ), ∀ Cq ⊂ T , (1.44)
where Cq refers to the dimension index in the related binary tree and X[q] is the
matricization according to Cq .
Let AK denote a set of tensors whose hierarchical rank is no more than K, i.e.,
AK = {A ∈ RI1 ×···×IN |rank(A[q] ) KCq , ∀ Cq ⊂ T}. Using the nestedness property,
we can define HT decomposition properly.
1.3 Tensor decompositions 21
×K
where UCq ∈ RICq ×KCq , VCq ∈ R C̄q Cq , KCq is the hierarchical rank of A[q] , and
I
Cq ⊂ T. Therefore, for Cq = {Cq1 , Cq2 }, the column vectors UCq (:, l) of UCq satisfy
the nestedness property when 1 l KCq , that is,
KCq KCq
1 2
UCq (:, l) = GCq (l, l1 , l2 )UCq1 (:, l1 ) ⊗ UCq2 (:, l2 ), (1.45)
l1 =1 l2 =1
where GCq (l, l1 , l2 ) represents the entry in GCq . It is the coefficient in the linear com-
bination of vectors. Here UCq1 (:, l1 ) and UCq2 (:, l2 ) are the column vectors of UCq1
and UCq2 . Therefore, GCq and UCq are the factors of HT decomposition of A. Ac-
cording to the dimension tree in Fig. 1.21, the HT decomposition for a fourth-order
tensor is illustrated in Fig. 1.22.
R1 RN+1
X= ··· G (1) (r1 , :, r2 ) ◦ G (2) (r2 , :, r3 ) ◦ · · · ◦ G (N ) (rN , :, rN +1 ), (1.46)
r1 =1 rN+1 =1
22 CHAPTER 1 TDs: computations, applications, and challenges
FIGURE 1.22
Hierarchical Tucker decomposition of a fourth-order tensor A ∈ RI1 ×I2 ×I3 ×I4 , where
G{1,2,3,4} ∈ RK{1,2} ×K{3,4} ×1 , G{1,2} ∈ RK{1,2} ×K{1} ×K{2} , G{3,4} ∈ RK{3,4} ×K{3} ×K{4} ,
U{1} ∈ RK{1} ×I1 , U{2} ∈ RK{2} ×I2 , U{3} ∈ RK{3} ×I3 , and U{4} ∈ RK{4} ×I4 are the factors of
hierarchical Tucker decomposition.
FIGURE 1.23
Top: Tensor train decomposition on an Nth-order tensor X . The yellow dot (light gray in
print version) G (n) ∈ RRn ×In ×Rn+1 is a core tensor, leaf components (not drawn) are
identities in this case and thus no need to be stored. Bottom: Corresponding core tensors.
or in an element-wise style,
xi1 ,i2 ,··· ,iN = G (1) (:, i1 , :)G (2) (:, i2 , :) · · · G (N ) (:, iN , :), (1.47)
where G (n) (n = 1, · · · , N) are the Rn × In × Rn+1 core factors. Note that we set
R1 = RN +1 = 1.
Definition 1.3.17. (Tensor train rank) The TT rank of X ∈ RI1 ×···×IN is an (N + 1)-
tuple
FIGURE 1.24
Tensor ring decomposition of an N-th-order tensor X , where G (n) ∈ RRn ×In ×Rn+1 are the
core tensors.
From Fig. 1.24, we can see that TR decomposes an N -th-order tensor into N third-
order factor tensors whose dimensions are much smaller than the original one. The
constraint R1 = RN +1 still exists but not necessarily equals one. Therefore, the defi-
nition of TR decomposition can be given.
Definition 1.3.18. (TR decomposition) The TR decomposition of X ∈ RI1 ×···×IN is
R1
RN
X= ··· G (1) (r1 , :, r2 ) ◦ G (2) (r2 , :, r3 ) ◦ · · · ◦ G (N ) (rN , :, r1 ). (1.49)
r1 =1 rN =1
CHAPTER X.
THE SIEGE.
CHAPTER XI.
CHAPTER XII.
THE ORPHAN.
CHAPTER XIII.
THE ARREST.
ebookmasss.com