Accepted Manuscript
Orthogonal Tubal Rank-1 Tensor Pursuit for Tensor Completion Weize Sun, Lei Huang, H.C. So, Jiajia Wang PII: DOI: Reference:
S0165-1684(18)30381-5 https://doi.org/10.1016/j.sigpro.2018.11.015 SIGPRO 6989
To appear in:
Signal Processing
Received date: Revised date: Accepted date:
14 November 2017 18 October 2018 20 November 2018
Please cite this article as: Weize Sun, Lei Huang, H.C. So, Jiajia Wang, Orthogonal Tubal Rank-1 Tensor Pursuit for Tensor Completion, Signal Processing (2018), doi: https://doi.org/10.1016/j.sigpro.2018.11.015
This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT
Highlights • A novel tensor completion method based on tensor tubal rank minimiza-
CR IP T
tion and orthogonal matching pursuit is proposed. • The convergence under the noise-free case is proved.
• The algorithm is modified to overcome the tubal impulsive noises by employing tensor tubal lp-norm optimization with 1 < p < 2.
• Experimental results show that the proposed methods outperform the
AC
CE
PT
ED
M
putational complexity.
AN US
state-of-the art algorithms in terms of RMSE, PSNR, SSIM and/or com-
1
ACCEPTED MANUSCRIPT
Orthogonal Tubal Rank-1 Tensor Pursuit for Tensor Completion
a College
CR IP T
Weize Suna , Lei Huanga , H.C. Sob,∗, Jiajia Wanga of Information Engineering, Shenzhen University, Shenzhen, Guangdong, China. of Electronic Engineering, City University of Hong Kong, Hong Kong
b Department
Abstract
AN US
This work addresses the issue of tensor completion. The properties of the tensor tubal rank are firstly discussed. It is shown that the tensor tubal rank has similar properties like that of matrix rank derived from SVD. The completion algorithm for the case that the measurements are noise-free or corrupted by Gaussian noise is then proposed based on an orthogonal pursuit on tubal rank-1 tensors. The philosophy behind the devised approach is to relax the problem of tensor tubal
M
rank minimization into tensor Frobenius-norm optimization with a constraint on the maximum number of orthogonal tensors. An iterative procedure which
ED
calculates one orthogonal tensor at each iterative step is then suggested, and the local convergence under the noise-free case is also proved. Furthermore, the proposed method is generalized to the situation where the observations are
PT
corrupted by impulsive noise in a tubal form. To tackle the impulsive noise, we formulate the problem of tensor completion as minimization of tensor tubal `p -norm with 1 < p < 2. An iteratively reweighted procedure is employed to
CE
compute the orthogonal tensors. The algorithms are compared with the stateof-the-art approaches using both synthetic data and real data sets.
AC
Keywords: Tensor completion, tensor SVD, tensor pursuit, rank-one tensor, tensor tubal rank.
∗ Corresponding
author work described in this paper was supported by the National Natural Science Foundation of China (NSFC) under Grant 61501300 and U1501253. 1 The
Preprint submitted to Elsevier
November 26, 2018
ACCEPTED MANUSCRIPT
1. Introduction Recovery of multi-dimensional data under limited number of measurements, which is referred to as matrix completion or tensor completion, is an important
CR IP T
problem which has drawn increasing attention in recent years. This problem is widely seen in various fields such as recommendation systems [1], dimensionality reduction [2], speech signal processing [3][4], MIMO radar [5][6], data mining [7][8], multi-class learning [9], and computer vision [10][11].
The idea of recovering unmeasured entries of a matrix basically relies on the
framework of finding a low-rank matrix that can model the original data, as
AN US
turns out to be the task of minimizing the matrix rank or the `0 -norm of the singular values of the matrix in the general case. However, it is a non-convex
NP-hard problem. One conventional solution is to minimize the matrix nuclear norm or the `1 -norm of the singular values instead of matrix rank, therefore relaxing the non-convex problem to a tractable and convex one [12][13][14]. Based
M
on this concept, a number of algorithms, such as Singular Value Thresholding (SVT) [15], Grassman Manifold based method [16], Singular Value Projection [17] and Fixed Point Continuation with Approximate SVD (FPCA) [18], have
ED
been proposed.
For tensors which are regarded as a generalization of vectors and matrices,
PT
although the process of data recovery is more natural as it employs the intrinsically multi-dimensional structure, the problem is more complicated to solve because the lack of a unique definition of tensor rank. One type of methods is to
CE
use the minimum number of rank-1 tensors from the CANDECOMP/PARAFAC (CP) decomposition [19][20][21]. However, the decomposition is very costly, and sometimes fails to give reasonable result because of the ill-posedness property
AC
[22]. Other researchers minimize the Tucker rank as the tensor rank from the Tucker decomposition to achieve good reconstruction results, but these algorithms require the knowledge of tensor rank [23] or a rank estimate [5] thereby being only valid for special applications. Moreover, some of the tensor completion methods based on the CP or Tucker rank minimization rely on the usage
3
ACCEPTED MANUSCRIPT
of optimization toolbox [24] thus leading to a very high complexity [25]. Other researchers, on the other hand, have proposed to solve it as a low-n-rank tensor recovery problem, which first unfolds an R-D tensor to R matrices in R ways
CR IP T
and then minimizes the sum of the ranks or nuclear norms of all the matrices for data recovery [26]. However, it may fail to exploit the tensor structure with the unfolding operations and thereby leads to a suboptimal procedure [27].
Furthermore, when the tensor is very large, these methods have to perform Singular Value Decomposition (SVD) on several large matrices at each iteration, resulting in considerably high computational complexity [28].
AN US
Recently, a new type of tensor decomposition based on tensor product is
proposed [29]. This factorization method first defines the tensor tubal rank as the tensor rank and then decomposes the tensor into several sub-tensors using the idea of unraveling the multi-dimensional structure by constructing group rings along the tensor tubes. The advantage of such an approach is that the resulting algebra and analysis are very close to those of matrix algebra and
M
analysis [30], and it is shown to be suitable for tensor completion [11]. However, similar to the matrix nuclear norm based methods [15], these algorithms require
ED
a selection of several user-defined parameters. With inappropriate parameters, a severe downgrade of performance will be resulted. Moreover, most existing methods cannot deal with the situation when there are outlier measurements
PT
[31][32]. In this work, we propose a novel tensor completion method based on tensor tubal rank [29][30] minimization as well as orthogonal matching pursuit
CE
[33]. A modification is then applied for tensor completion under tubal impulsive noises by employing tensor tubal `p -norm optimization on the residual between
AC
the measurements and recovered data with p < 2. The rest of the paper is organized as follows. In Section 2, the notations
and required definitions are first presented. In Section 3, our tensor recovery approaches for both random sampling and tubal sampling cases are developed, and the convergence is also proved. In Section 4, a modified algorithm is devised to tackle the situation when the measured entries are corrupted by tubal outliers. Finally, numerical results are provided in Section 5, and conclusions are drawn 4
ACCEPTED MANUSCRIPT
in Section 6.
2. Notations, Definitions and Preliminaries
CR IP T
Before going to the main result, we present the notations and preliminaries
in this paper. We mainly follow the definitions in [29][30], and some notations are newly defined in this section.
Scalars, vectors, matrices and tensors are denoted by italic, bold lower-case, bold upper-case and bold calligraphic symbols, respectively. The transpose and conjugate transpose of a vector or a matrix are written as T and H , and the i × i
AN US
identity matrix is symbolized as Ii . To refer to the (m1 , m2 , . . . , mR ) entry of an
A}(m1 ,m2 ,...,mR ) . The element-wise R-D tensor A ∈ CM1 ×M2 ×···×MR , we use {A multiplication operator between two tensors is defined as , while the symbol t represents the concatenation operator where A = A 1 tr A 2 is obtained by stacking A 2 to the end of A 1 along the rth dimension. C
M
For a compact notation, the m3 th frontal slice of a 3-D tensor A ∈
M1 ×M2 ×M3
is defined as A {m3 } = A (:, :, m3 ) for m3 = 1, 2, · · · , M3 , that
is, A = A {1} t3 A {2} t3 · · · t3 A {M3 } , and an M1 × M1 × M3 identity tensor is
ED
{1} denoted by I M1 M1 M3 with I M1 M1 M3 = IM1 and all the other entries being 0.
Furthermore, the tensor conjugate transpose of a 3-D tensor A ∈ CM1 ×M2 ×M3
PT
is an M2 × M1 × M3 tensor A H which is obtained by taking the conjugate transpose of each of the frontal slices and then reversing the order from the second through the M3 th conjugate transposed frontal slices.
CE
The Fast Fourier Transform (FFT) along the third dimension of all the tubes
e = fft(A A, [ ], 3). Therefore, in the same way, A can of a tensor A is denoted as A
be computed by applying inverse FFT (IFFT) along the third dimension of all
AC
e . We can then define the t-product [29] between two tensors: the tubes of A
Definition 2.1. t-product: The t-product ∗ between two tensors A ∈
RM1 ×M4 ×M3 and B ∈ RM4 ×M2 ×M3 is defined as
C = A ∗ B ∈ RM1 ×M2 ×M3
5
(1)
ACCEPTED MANUSCRIPT
where {m3 } e {m3 } e {m3 }B Ce =A
(2)
CR IP T
for m3 = 1, 2, · · · , M3 , that is, the actual computation of C can be divided into
3 steps: FFT along the third dimension of A and B , matrix multiplications, and
IFFT along the third dimension of Ce. Now the orthogonal tensor is defined as
follows.
Definition 2.2. Orthogonal tensor: A tensor Q ∈ RM1 ×M1 ×M3 is called
AN US
orthogonal if Q H ∗ Q = Q ∗ Q H = I M1 M1 M3 .
(3)
Furthermore, a tensor Q ∈ RM1 ×M2 ×M3 , with M1 ≤ M2 , is called left orthogonal if Q H ∗ Q = I M1 M1 M3 .
Definition 2.3. Block diagonal form: Let A denote the block diagonal
e: matrix of tensor A
M
h {1} e ) = blockdiag( A e A , blockdiag(A
ED
e {2} A
···
i e {M3 } ) ∈ CM1 M3 ×M2 M3 A
(4)
According to this definition, we can see that the block diagonal matrix of A H
PT
is equal to the conjugate transpose of the block diagonal matrix A [30]:
CE
Definition 2.4.
H
AH = A .
(5)
Tensor Singular Value Decomposition (t-SVD):
Analogous to the SVD of a 2-D matrix, the t-SVD of a 3-D tensor X ∈
AC
RM1 ×M2 ×M3 is written as [29]
X = U ∗ S ∗ VH =
Ms X
m=1
Um ∗ Sm ∗ VH m
(6)
where Ms = min{M1 , M2 }, U = U 1 t2 U 2 t2 · · · t2 U Ms ∈ RM1 ×Ms ×M3 , V =
V 1 t2 V 2 t2 · · · t2 V Ms ∈ RM2 ×Ms ×M3 , U m = U (:, m, :) and V m = V (:, m, :) for m = 1, 2, . . . , Ms . The S ∈ RM1 ×M2 ×M3 is defined as a rectangular f-diagonal 6
ACCEPTED MANUSCRIPT
tensor with all frontal slices being diagonal matrices. Note that U and V are
orthogonal tensors. Here we name S m ∈ R1×1×M3 as the mth singular value
V m ) as the left (right) singular vector slices. Actually, tube of tensor X and U m (V
CR IP T
this decomposition can be obtained by applying matrix SVD on all the frontal e to get Ue , Se and Ve and then performing IFFT on them [29]. slices of X
Based on t-SVD, one can now define the tensor ranks [29] as follows.
Definition 2.5. Tensor multi-rank and tensor tubal rank (TTR):
The tensor multi-rank of a tensor X ∈ RM1 ×M2 ×M3 is a vector r ∈ ZM3 ×1 with e . The TTR the m3 th element indicating the rank of the m3 th frontal slice of X
AN US
X ), is defined as the number of non-zero singular value of X , written as TTR(X tubes of S , which is r = max{r}. If a third-order tensor X is of full TTR, it means that r = min{M1 , M2 }.
Definition 2.6. Inner product of tensors: If A and B are third-order tensors of dimensions M1 × M2 × M3 , the inner product between A and B is defined as:
A}(m1 ,m2 ,m3 ) {B B }(m1 ,m2 ,m3 ) {A
(7)
m1 =1 m2 =1 m3 =1
ED
and
M1 X M2 X M3 X
M
< A , B >,
< A , B >=
1 1 H B A) = trace(B < A, B > . M3 M3
(8)
PT
Note that according to the conjugate symmetric property of FFT and (4), it can be easily verified.
CE
Definition 2.7. Tensor Frobenius-norm: The Frobenius-norm of a tenqP PM2 PM3 M1 A||F = A}2(m1 ,m2 ,m3 ) , has the folsor A , defined as ||A m1 =1 m2 =1 m3 =1 {A
AC
lowing relation according to the above definition A||F =< A , A >1/2 = √ ||A
1 e ||F = √ 1 ||A A||F . ||A M3 M3
(9)
Remark 2.1. According to these definitions, by defining the normalized
A} = A /||A A||F we get tensor of A ∈ RM1 ×M2 ×M3 as normalize{A e} = normalize{A
p
7
M3
e /||A e ||F . A
(10)
ACCEPTED MANUSCRIPT
Definition 2.8. Tensor mode-2 multiplication: The tensor mode-2 e ∈ CM1 ×M4 ×M3 and B e ∈ CM4 ×M2 ×M3 is multiplication between two tensors A
e ?B e ∈ CM1 ×M2 ×M3 , where Ce defined as Ce = A
{m3 }
{m3 }
e =A
e {m3 } , which in turn B
e and B e. A
The novel tubal `p -norm is defined as follows.
CR IP T
means that each frontal slice of Ce is a product of corresponding frontal slices of Definition 2.9. Tensor tubal `p -norm: The tensor tubal `p -norm of a
tensor A is computed by first calculating an l2 -norm of all its tubes and then computing the `p -norm of the resulting matrix: M1 X M2 X
AN US
A||t,p = ||A||t,p = ( ||A
{A}p(m1 ,m2 ) )1/p
(11)
m1 =1 m2 =1
A(m1 , m2 , :)||2 or A , ||A A||2,: . Here, A (m1 , m2 , :) is the where {A}(m1 ,m2 ) = ||A (m1 , m2 )th tube of A . Note that (11) can be taken as a tensor form of the `2,p -norm [34].
M
Remark 2.2. According to Definition 2.9, we have
ED
A||t,p = √ ||A
1 e ||t,p . ||A M3
(12)
Remark 2.3. The tensor tubal `2 -norm of a tensor is its Frobenius-norm.
PT
3. Main Result
In this section, we derive the tensor recovery algorithm using 3-D data as an
CE
example. Note that for higher-dimensional data, we can recursively apply the Fourier transform over successive dimensions higher than three [35] and then
AC
unfold the whole tensor into a 3-D one. 3.1. Rank-1 Tensor Pursuit with Random Sampling We first formally formulate the tensor completion problem with random
sampling, which means that any of the measurements in a tensor Y = S + N ∈
RM1 ×M2 ×M3 is sampled independently with a probability q1 ∈ (0, 1), while S
and N refer to the signal and noise terms. The q1 is also named as observation 8
ACCEPTED MANUSCRIPT
percentage. Our objective is to recover S from Y Ω , namely, the observed entries of Y where Ω is an M1 × M2 × M3 tensor with elements being 1 or 0 indicating whether the corresponding entries are observed or not, respectively. Generally
CR IP T
speaking, this problem can be formulated as X) min rank(X
(13)
X
s.t. X Ω = S Ω .
(14)
The tensor rank can be described using the CP [21] or Tucker [23] criterion, or P3 relaxed to a summation of several matrix ranks as i=1 αi rank(Xi ) where Xi
AN US
is the i-way unfolding of the tensor X [26][27]. In this work, we assume that the tensor S has a limited tubal rank r < min{M1 , M2 }. Then it is easy to verify that this tensor can be written as a linear combination of several tubal rank-1 tensors:
M
S = M (l) (θ (l) ) =
X
l∈{L}
θl M l
(15)
Ml : l ∈ {L}} is the set of all the tubal rank-1 tensors with unit where {M
ED
tensor Frobenius-norm, and θl is the magnitude of the l-th rank-1 tensor. The low tubal rank tensor completion problem can then be transformed into the
min ||θ||0 θ
M(θ))Ω = Y Ω = S Ω . s.t. (M
(16) (17)
CE
PT
following constrained minimization:
M(θ))Ω − Y Ω || ≤ where is a small For the noisy scenarios, (17) becomes ||(M value. It is worth noting that by relaxing the `0 -norm to `1 -norm, the resultant
AC
problem is identical to minimizing the tensor nuclear norm (TNN) [30]. We can then reformulate the problem as Y Ω − (M M(θ))Ω ||2F min ||Y
(18)
s.t. ||θ||0 ≤ R
(19)
θ
9
ACCEPTED MANUSCRIPT
where R indicates the maximum number of tubal rank. According to (9), (18) is equivalent to Y Ω − (M M(θ))Ω ||2F min ||Y
(20)
CR IP T
θ
s.t. ||θ||0 ≤ R
(21)
which can be solved by the orthogonal matching pursuit. In the kth iteration, the tubal rank-1 basis tensors M 1 , M 2 , · · · , M k−1 and their weights θ (k−1) have
already been computed in the former (k−1) iterations. The recovered tensor and
X (k−1) )Ω , the residual term are X (k−1) = M (k−1) (θ (k−1) ) and R (k) = Y Ω − (X
AN US
respectively. When R (k) = 0, which is a zero matrix, the problem is solved and
the rank of the recovered tensor is at most k − 1. On the other hand, for a nonzero R (k) , the next M k can be solved by
Mk ||F = 1} max{< M k , R (k) >: M k = U k ∗ S k ∗ V H k , ||M Mk
(22)
M
where U k , S k , V k are tensors of the dimensions M1 × 1 × M3 , 1 × 1 × M3 , M2 × 1 × M3 , respectively. According to (4), the optimization of the block
ED
diagonal term is substituted by a series of optimization of each block: gk max {< M
gk {m3 } M
{m3 }
g (k) ,R
{m3 }
gk >: M
{m3 }
= Ufk
{m3 }
Sfk
{m3 }
H
(Vfk ){m3 } } (23)
Ufk
PT
Mk ||F = 1. Then we obtain the for m3 = 1, 2, · · · , M3 under the constraint ||M {m3 }
H
gk and (Vfk ){m3 } as the top left and right singular vectors of M
[33] which can be solved using the power method [36] [37], and Sfk
CE
gk corresponding maximum singular value. Once all the M
{m3 }
{m3 }
{m3 }
is the
are computed,
AC
the normalization process of M k can be carried out according to (10) or gk ← M
p
gk /||Sfk ||F M3 M
(24)
and the norm of M k is thus restricted to 1.
Then the new θ (k) is determined by solving the least squares (LS) problem: ˙ (k) θ (k) ||2 min ||y˙ − M θ (k)
10
(25)
ACCEPTED MANUSCRIPT
h i ˙ (k) = m Y Ω }, M Mi )Ω } for ˙ i = vec{(M where y˙ = vec{Y ˙ 1, m ˙ 2. · · · , m ˙ k and m i = 1, 2, · · · , k.
The above two steps can run iteratively until a stopping criterion is satisfied,
CR IP T
and this Tubal Rank-1 Tensor Pursuit (TR1TP) algorithm is summarized in
Algorithm 1. Furthermore, we prove that the algorithm converges and achieves a linear convergence rate in Appendix A, which satisfies:
R(k+1) ||22 ≤ (1 − 1/ min{M1 , M2 })||R Rs (k)||22 ≤ (1 − 1/ min{M1 , M2 })k ||Y Y Ω ||22 . ||R
(26)
Step 1: Initialize R (1) = Y Ω .
AN US
Algorithm 1: Tubal Rank-1 Tensor Pursuit Algorithm for Tensor Completion
g (k) . gi , i = 1, 2, · · · , k − 1, calculate its Fourier domain R Step 2: Given R (k) and M for m3 = 1, 2, · · · , M3
Step 3: Find a pair of top left and right singular vectors (Ufk gk set M
{m3 }
end
= Ufk
{m3 }
M
corresponding singular value Sfk
{m3 }
Sfk
{m3 }
Vfk
{m3 }
g (k) of R
{m3 }
{m3 }
, Vfk
{m3 }
) and its
using the power method, and
.
ED
gk according to (24). Step 4: Normalize M
Step 5: Compute θ (k) as the LS solution of (25).
PT
X (k) )Ω . Step 6: Assign X (k) = M (k) (θ (k) ) and update R (k+1) = Y Ω − (X Step 7: Repeat Steps 2 to 6 until the maximum iteration number R, or
CE
X (k) − X (k−1) ||2 /||X X (k−1) ||2 < , is reached, and output X . ||X Note that this algorithm has only one user-defined parameter, which is the
maximum number of iterations or the tubal rank of the tensor, denoted as R.
AC
However, when the number of iterations is large, the LS solution of (25) needs to compute the inverse of a large matrix, making it computationally unattractive. To overcome this problem, an economic updating scheme which updates α = h i α1 , α2 instead of θ in Algorithm 1 is suggested, and we solve ˙ k α2 ||2 min ||y˙ − x˙ (k−1) α1 − m αk
11
(27)
ACCEPTED MANUSCRIPT
X (k−1) )Ω } and X (k) = α1X (k−1) + instead of (25) by assigning x˙ (k−1) = vec{(X α2M k . As the convergence analysis of the economic updating scheme is similar
CR IP T
to that of Algorithm 1, it is ignored herein. 3.2. Rank-1 Tensor Pursuit with Random Tubal Sampling
Sometimes the received data are sampled in tubal format [30] [38], which means that any one of M1 M2 tubes along the third dimension of a tensor Y ∈
RM1 ×M2 ×M3 is recorded with a probability q1 independently. In this case, if the Y }(m1 ,m2 ,m3 ) is recorded, then all the entries along the (m1 , m2 ) tube entry {Y
AN US
are available. Therefore, it is called ‘random tubal sampling’, and a sampling matrix Ω of dimensions M1 × M2 with elements being 1 or 0 is used to refer to the measured and non-measured tubes of Y instead of Ω in (17). However, for consistency, we follow the notation Y Ω as the observed entries tubes of Y like that in previous sections.
It is easy to see that the Fourier transform operation along the third dimen-
M
sion does not change the structure of the observed entries, therefore we can use e Ω to represent the observed entries of Y e and rewrite (20) as the notation Y
ED
e Ω − (M f (θ))Ω ||2 min ||Y F θ
(28)
which indicates that in this case, we only need to perform the FFT once and
PT
then finish all the computations in the Fourier domain. The modified method, named as the TR1TP for random Tubal sampling (TR1TPT), is shown below
CE
as Algorithm 2. Following the same techniques in Appendix A, Algorithm 2 can also be proved to achieve a linear convergence rate. For compactness of
AC
presentation, it is not shown here.
4. Rank-1 Tensor Pursuit with Random Tubal Outlier In this section, we focus on the problem that the observed data are corrupted
by tubal impulsive noise, that is, some tubes of a tensor Y ∈ RM1 ×M2 ×M3 are
corrupted by impulsive noise while the other tubes are clean or only added by
12
ACCEPTED MANUSCRIPT
Algorithm 2: TR1TPT Algorithm for Tensor Completion. (1)
e Step 2: Initialize R
for m3 = 1, 2, · · · , M3
eΩ . =Y
CR IP T
e. Step 1: Calculate Fourier transform of Y as Y
e (k) and M gi , i = 1, 2, · · · , k − 1, find a pair of top left Step 3: Given R and right singular vectors (Ufk
singular value Sfk
gk set M
{m3 }
end
{m3 }
= Ufk
{m3 }
{m3 }
, Vfk
{m3 }
) and its corresponding
e (k−1) ){m3 } using the power method, and of (R
Sfk
{m3 }
(Vfk
{m3 }
)H .
(k)
e (k) = M f Step 6: Assign X
AN US
gk according to (24). Step 4: Normalize M h i ˙ †˙ ˙ f f ˙ ˙ ˙ e Step 5: Compute θ (k) as θ (k) = M y , where M = e e e m , m . · · · , m k k 1 2 k .
e (k+1) = Y e Ω − (X e (k) )Ω . (θ (k) ) and update R
Step 7: Repeat Steps 3 to 6 until the maximum iteration number R, or e (k) − X e (k−1) ||2 /||X e (k−1) ||2 < , is reached, and output the IFFT ||X
M
e. of X
additive Gaussian noise. This phenomenon has been well elaborated in [32][39].
ED
For ease of presentation, we assume that the observed tensor is sampled by
PT
random tubal sampling, and define the additive noise as E 1 + q2E 2 E = (1 − q2 )E
(29)
where the entries in E 1 are zero or follows a zero mean white Gaussian distri-
CE
bution, and E 2 are outlier tubes. The q2 ∈ (0, 1) in (29) is the probability that the (m1 , m2 ) tube of E is sampled from E 1 and E 2 with probabilities of 1 − q2
AC
and q2 , respectively. The observed data become S + E )Ω Y Ω = (S
(30)
where S has a low tensor tubal rank, and the target is to recover a low-rank tensor X → S from Y Ω . When E 1 = 0, (30) can be formulated as a minimization
13
ACCEPTED MANUSCRIPT
of tubal `0 -norm Y Ω − X Ω ||t,0 min ||Y
(31)
s.t. ||θ||0 ≤ R,
X = M (θ)
CR IP T
θ
where the tubal `0 -norm is used to indicate the location of outliers. However, this is an NP hard problem and is not suitable for the more general case that E 1 6= 0 . Therefore, we relax (31) to tubal `p -norm for p ∈ (1, 2): Y Ω − (M M(θ))Ω ||t,p min f = ||Y θ
(32)
AN US
which can also be solved by utilizing the matching pursuit type algorithm. Note
that when the data are sampled by random tubal sampling and corrupted by random tubal outliers, the completion can be carried out in Fourier domain as in Section 3.2. Here we follow the procedure of economic pursuit method. The equations of data recovery under random sampling and/or using the full weighting vector θ are similar and thus not shown here.
M
e (k−1) In the kth iteration, the recovered tensor and the residual term are X
e (k) = Y eΩ − X e (k−1) , respectively. Subsequently, we relax the Fourier and R Ω
ED
formulation of (32) to a two-step problem that first solves min
R(k) − (U Uk ∗ VH f = ||R k )Ω ||t,p
PT
U k ∗V VH M k =U k
(33)
where U k and V k are tensors of the dimensions M1 × 1 × M3 and 1 × M2 × M3 .
CE
According to (12), (33) is equivalent to min
g H ) f k =(U e k ?V M k Ω
g H ) || e (k) − (U ek ? V f = ||R k Ω t,p
(34)
AC
and then we can solve the problem using an alternately block updating apg H under a given U e k and then Ue k+1 with a fixed proach, that is, updating V k
g H iteratively. Then we can compute V k min
α k =[α1 ,α2 ]
e Ω − α1 (X e (k−1) )Ω − α2 (M f k )Ω ||t,p ||Y
f k according to (24). after normalizing M
14
(35)
ACCEPTED MANUSCRIPT
Assume that in the i-th iteration, the (Ue k )(i) is already computed, and the g g H are both of tubal rankH . As long as U e k and V target is now to determine V k k
g H = H t H t ···t H e = Z 1 t2 Z 2 t2 · · · t2 Z M and V 1, we can write R k 1 2 2 2 2 M2 2
Z m2 − G m2 ? H m2 ||t,p min ||Z
H m2
CR IP T
e where Z m2 ∈ RM1 ×1×M3 and H m2 ∈ R1×1×M3 are the m2 -th slice or tube of R g H , respectively, and change (34) into solving and V k
(36)
e k Ω m and Ω = Ω 1 t2 Ω 2 t2 · · · t2 Ω M with Ω m being the where G m2 = U 2 2 2
same size of Ue k with values 0 or 1. Noting that in (34), the missing tubes of
AN US
Z m2 and G m2 are the same, and we can use the tubes that are only observed, thus reducing the data size and computational complexity.
W ||2,: ∈ RM1 ×1 , we realize By defining W = Z m2 − G m2 ? H m2 and W = ||W that the minimization problem becomes
H m2
M1 X
(W)p(m1 ) =
m1 =1
2 (W)p−2 (m1 ) (W)(m1 )
m1 =1
W||22
D W ||22 = ||D D Z m2 − (D D G m2 ) ? H m2 ||2t,2 = ||D (37)
ED
= ||D
M1 X
M
min ||W||pp =
where D {m3 } = D = W(p−2)/2 for m3 = 1, 2, · · · , M3 . According to (4), (9)
PT
and defining Z D = D Z m2 as well as G D = D G m2 , (36) becomes Z D ) − blockdiag(G GD ) min ||blockdiag(Z
Hm2
H m2 )||22 blockdiag(H
(38)
CE
which is a matrix based optimization problem that can be solved by LS fit. {m } {m3 } † {m3 } GD Therefore the m3 -th slice of H m2 is calculated as H m2 3 = (G ) Z D . The g H can now be computed using such as the Tubal Iteratively Reweighted Least V k
AC
g H and D iteratively until convergence. Squares (TIRLS) method by updating V k (i) (i+1) g g H H can be retrieved following the similar Once (V k ) is obtained, (U k ) procedure of (34) - (38), and terminates when a stopping criterion is reached.
We are now at a position to solve (35). To this end, by defining W α =
{m3 } (p−2)/2 e Ω − α1 (X e (k−1) )Ω − α2 (M f k )Ω , Wα = ||W W α ||2,: and D α Y = Dα = Wα
15
ACCEPTED MANUSCRIPT
for m3 = 1, 2, · · · , M3 , (35) becomes min
α (k) =[α1 ,α2 ]
e )Ω − α1 (D e k−1 )Ω − α2 (D e k )Ω ||2 Dα Y Dα X Dα X ||(D
(39)
CR IP T
e )Ω , (D e k−1 )Ω Dα Y Dα X and then α (k) is calculated using LS after vectorizing (D e k )Ω . Dα X and (D
α(k) )ini = [1, 1], it is updated iteratively with With an initialization of (α
{m } D α 3 until a stopping criterion is satisfied. The proposed TIRLS based eco-
nomic TR1TPT algorithm (TIRLS-TR1TPT-Econ) is summarized as Algorithm
AN US
3.
Algorithm 3: TIRLS-TR1TPT-Econ Algorithm for Tensor Completion e and set R e (1) = Y eΩ . Step 1: Calculate Fourier transform of Y as Y (k) (k−1) g H following e e Step 2: Given R and X , calculate Ufk and V k g H . gk = Ufk ? V (34) - (38) and set M k
gk according to (24). Step 3: Normalize M e Step 5: Assign X
(k)
M
Step 4: Compute α (k) according to (39). e = α1X
(k−1)
f k and update R e + α2M
(k+1)
e Ω − (X e =Y
(k)
)Ω .
Step 6: Repeat Steps 2 to 4 until the maximum iteration number R, or
PT
e. of X
ED
e (k) − X e (k−1) ||2 /||X e (k−1) ||2 < , is reached, and output the IFFT ||X
CE
5. Numerical Results In this section, we show the numerical results, based on the synthetic data as
well as real image and video data, of the proposed algorithms and state-of-the-
AC
art algorithms. The first subsection considers the standard tensor completion problem while the second one tackles the problem in the presence of tubal outliers. All the experiments are repeated 100 times to obtain the average performance.
16
ACCEPTED MANUSCRIPT
5.1. Tensor Completion under Random Tubal Sampling and Random Sampling In the first test, we study the convergence of the proposed ‘TR1TP’ and ‘TR1TPT’ methods as well as their economic variants, named as ‘TR1TP-Econ’
CR IP T
and ‘TR1TPT-Econ’. The element-wise Root Mean Square Error (RMSE), deX −S S ||F /||S S ||F , where X and S are the recovered and original tensors, fined as ||X is used to evaluate the performance of data recovery. The dimensions of the synthetic data tensor S is 100 × 100 × 30 of tubal rank 5, and its elements are
sampled under the rules of random tubal sampling with two different sampling
rates q1 = 40% and 80%. The empirical RMSEs versus the iteration number
AN US
R in noise-free and Gaussian noise situations are plotted in Figures 1 and 2, respectively. Note that, for the noisy case, the observed entry tubes are corrupted by an additive white Gaussian noise (AWGN) and the signal-to-noise ratio (SNR) is 30dB. It is shown that for all the methods, as the number of
iterations increases, the recovery performance improves. When R is larger than a certain threshold R0 , such as R > R0 = 30 at q1 = 40% under noise-free
M
case, the proposed algorithms converge, and have good performance for a wide range of R. Therefore, we can simply set R to a large number to get acceptable
ED
results and avoid performance downgrade. For the noisy scenario, after several iterations, the performance of the proposed algorithms degrades slightly as the number of iterations increases, showing that in this case, the new recovered
PT
results from new iterations are actually noise. It is worth noting that for both the noise-free and noisy cases, as the sampling rate increases, the number of
CE
iterations required to converge increases, indicating that R should be chosen in proportional to q1 . Furthermore, due to the de-noising effect of the FFT operation along the 3rd dimension, the recovery performance of ‘TR1TPT’ and
AC
‘TR1TPT-Econ’ methods are slightly better than ‘TR1TPT’ and ‘TR1TPTEcon’, showing the attractiveness of the former tensor completion technique under the tubal sampling model. In the second test, the proposed methods are compared with the Tensor
Nuclear Norm algorithm based on t-SVD (‘t-SVD-TNN’) [30], the economic Rank-1 Tensor Pursuit (‘R1TP-Econ’) method [40] as well as the Fixed Point 17
ACCEPTED MANUSCRIPT
-1
0 TR1TP TR1TP-Econ TR1TPT TR1TPT-Econ
-5
-4
RMSE (dB)
RMSE (dB)
-3
-5 -6
CR IP T
-2
-10
-15
-7 -20 -8
0
20
40
Number of iterations
-25
AN US
-9
60
0
20
40
60
Number of iterations
Figure 1: Average RMSE versus iteration number in noise-free case. Left: q1 = 40%; Right: q1 = 80%.
M
Low-Rank Tensor Recovery (‘FP-LRTC’) approach [26]. The random sampling model is employed thus only the ‘TR1TP’ and ‘TR1TP-Econ’ approaches are applicable. We change the parameter M2 from 80 to 180 and set q1 = 90%, with
ED
the other parameters being the same as those in Figure 1. All the algorithms will terminate when the normalized difference between adjacent iterations is small
PT
X (k) − X (k−1) ||2 /||X X (k−1) ||2 < where = 10−6 , or the N th enough, that is, ||X iteration is reached. According to Figure 1, we set N = R = 60 for the proposed and ‘R1TP-Econ’ approaches. For the other algorithms, N = 100 is used. For
CE
the t-SVD-TNN approach, the shrinkage parameter is set to be τ = 0.025P ,
Y Ω ||2 is the `2 -norm of the observed entries, and decreases by a where P = ||Y
AC
factor 0.95 after each iteration. The results of RMSE and computational time in seconds are shown in Figures 3 and 4. In this test, the ‘t-SVD-TNN’ algorithm gives the best performance for small data size, but it is much more complicated than the proposed ‘TR1TP-Econ’ one. The two proposed algorithms give the best performance when the size of tensor is sufficiently large, verifying the efficiency of the economic process according to Figure 4. On the other hand, the
18
ACCEPTED MANUSCRIPT
-1
0 TR1TP TR1TP-Econ TR1TPT TR1TPT-Econ
-2
-4
-5 -6
-8 -10 -12
-7
-14
-8
-16
0
10
20
30
Number of iterations
-18
AN US
-9
CR IP T
-6
-4
RMSE (dB)
RMSE (dB)
-3
-2
40
0
10
20
30
40
Number of iterations
Figure 2: Average RMSE versus iteration number in Gaussian noise. Left: q1 = 40%; Right: q1 = 80%.
to the modeling error.
M
‘R1TP-Econ’ and ‘FP-LRTC’ methods provide unsatisfied results, partly due
The effect of Gaussian noise on the recovery performance is now examined.
ED
The empirical RMSEs of various techniques versus SNR are shown in Figure 5. We employ the same parameters as those in Figure 2 except q1 = 70%
PT
and R = 25. It is observed that the ‘t-SVD-TNN’ and proposed methods give similar recovery performance, and they can retrieve the data tensor correctly when the SNR is sufficiently large. The average computational times
CE
of the ‘TR1TP’, ‘TR1TP-Econ’, ‘t-SVD-TNN’, ‘R1TP-Econ’ and ‘FP-LRTC’ are 5.1025s, 1.4534s, 11.7652s, 1.4645s and 5.5055s, respectively, indicating that
AC
the ‘TR1TP-Econ’ is the most computationally attractive one among all the methods. To evaluate the applicability of real-world problems, we test our methods
and other state-of-the-art algorithms on color image inpainting, that is, to fill in the missing pixel values of a partially observed image. We employ the peak SNR (PSNR), structural similarity index (SSIM) [41] and computational time to
19
ACCEPTED MANUSCRIPT
-5
-10
TR1TP TR1TP-Econ t-SVD-TNN R1TP-Econ FP-LRCT
-20
-25
-30
-35
90
100
110
120
130
140
150
160
170
180
AN US
-40 80
CR IP T
RMSE (dB)
-15
M2
Figure 3: Average RMSE versus M2 .
evaluate these approaches, and use the Berkeley Segmentation database [42] in
M
this experiment. This database contains a wide variety of natural scenes and has a total number of 100 test images with dimensions 321×481×3. For each image,
ED
we discard the last column and row to get the data tensor Y 1 ∈ [0, 255]320×480×3 , reshape it to a higher-order tensor of size 16 × 20 × 20 × 24 × 3, and then unfold the 3rd and 4th dimensions into the 1st and 2nd ones to form a new tensor Y 2
PT
of the same dimensions as Y 1 . This Y 2 can be regarded as a ’detail-background’ representation of the original image Y 1 [43], and the final inpainting results of all the algorithms are the average of recovered values of both tensors.
CE
The image recovery performance in noise-free environment is studied at first
and the random tubal sampling model is used. The proposed ‘R1TPT-Econ’
AC
algorithm is employed and compared with ‘t-SVD-TNN’, ‘R1TP-Econ’ and ‘FPLRTC’ methods at 20% observed tubes of all 100 images. The maximum number of iterations is N = R = 100 for all the algorithms, and the other parameters are the same as those in the previous test. The PSNR results of all the images are shown in Figure 6. The average PSNRs of ‘R1TPT-Econ’, ‘t-SVD-TNN’ and ‘R1TP-Econ’ methods are 22.0965dB, 22.1316dB and 21.5329dB, and the 20
ACCEPTED MANUSCRIPT
45
35
Time (s)
30 25 20 15 10 5
90
100
110
120
130
140
150
160
170
180
AN US
0 80
CR IP T
TR1TP TR1TP-Econ t-SVD-TNN R1TP-Econ FP-LRCT
40
M2
Figure 4: Average computational time versus M2 .
corresponding average SSIM values are 0.8991, 0.8970 and 0.8848, verifying that
M
the ‘t-SVD-TNN’ and proposed methods give the best image recovery performance. The ‘FP-LRTC’ approach, on the other hand, fails to give reasonable
ED
results.
In the second experiment of image inpainting, we test the recovery performance in the presence of noise. A small disturbance or noise N is added to
PT
the observed entries and N follows an independent and identically uniform distribution between [−25, 25]. Note that the values of noise-free tensor S vary from 0 to 255. The parameters are the same as those in previous test except
CE
the sampling rate q1 varies from 5% to 35%. The average PSNR and SSIM results of all the images are shown in Figure 7, and one typical results of 5 im-
AC
ages, namely, ‘Doll’, ‘Elephant’, ‘Dock’, ‘Rocks’ and ‘Landscape’, at q1 = 15% are shown in Figure 8. The average computational times of ‘TR1TPT-Econ’, ‘t-SVD-TNN’, ‘R1TP-Econ’ and ‘FP-LRTC’ are 43.97s, 86.47s, 18.17s, 0.9618s and 505.21s, respectively. It is shown that for small observation percentage, the ‘R1TPT-Econ’ approach gives the best performance with moderate complexities comparing to the state-of-the-art ones, indicating the robustness of the 21
ACCEPTED MANUSCRIPT
0 -2
CR IP T
RMSE (dB)
-4 -6 -8 -10
TR1TP TR1TP-Econ t-SVD-TNN R1TP-Econ FP-LRCT
-12
-16
0
5
10
AN US
-14
15
20
25
30
SNR (dB)
Figure 5: Average RMSE versus SNR
M
proposed methods against this severe environment.
5.2. Tensor Completion under Random Tubal Outliers In this section, the proposed ‘TIRLS-TR1TPT-Econ’ approach is compared
ED
with the state-of-the-art algorithms under the situation where the observed random tubal sampled data are corrupted by tubal impulsive noise. The methods
PT
to be compared with are the ‘IRLS-t-SVD’ [44], ‘IRLS-PARAFAC’ [45], and ‘t-SVD-TNN’ based on `2 -norm minimization. At first, we compare the completion performances of various methods under
CE
different outliers levels for synthetic data, which is a 3-D tensor of dimensions 100 × 100 × 20 and tubal rank 5. Only 60% of the tubal entries are observed,
AC
and 10% of the observed entries are corrupted by tubal outliers. The outliers
E 2 in (29) are generated using a zero mean Gaussian distribution as well as
S ||22 /||E E 2 ||22 ranging uniform distribution, with a signal-to-outlier ratio (SOR) ||S from −10dB to 10dB. For all the IRLS based methods, the p in `p -norm minimization is set to 1.5. The maximum number of iterations is N = R = 20 for the ‘IRLS-t-SVD’ and proposed methods and N = 100 for the others. As
22
ACCEPTED MANUSCRIPT
TR1TP-Econ t-SVD-TNN R1TP-Econ FP-LRCT
30
20 15 10 5 0
0
5
10
15
20
25
30
35
Image ID 30
PSNR(dB)
25 20 15 10
0 50
55
60
65
40
45
AN US
5
CR IP T
PSNR(dB)
25
70
75
80
85
90
95
50
100
Image ID
Figure 6: PSNR results of all 100 images.
M
the ‘IRLS-t-SVD’ approach is a hard threshold method which requires prior knowledge of the tubal rank of the data, the real tubal rank 5 is chosen. The remaining parameters are the same as those in Figure 5. The mean and stan-
ED
dard deviation (std) of RMSE results are shown in Table 1. It is shown that the proposed approach gives the best completion performance. When the SOR is large, the l2 -norm based ‘t-SVD-TNN’ method can achieve an acceptable recov-
PT
ery result, but it fails when the SOR is small. The PARAFAC-type approach, however, fails due to the modeling error.
CE
The images in Berkeley Segmentation database are now employed for image inpainting in the presence of outliers. The images are sampled tubal-wised at q1 = 40% and 60%, and q2 = 10% of the measured tubes are corrupted by salt-
AC
and-pepper noises. The PSNR and SSIM results of the proposed algorithm with p = 1.7 under different numbers of iterations or tensor components R are shown
in Figure 9 to illustrate the convergence of the methodology. Similar to the cases in Figures 1 and 2, it is observed that for a larger observation percentage q1 , a larger R is required. Nevertheless, a smaller R is suggested for tensor
23
ACCEPTED MANUSCRIPT
24 0.9 22
0.85
CR IP T
0.8
20
18
SSIM
PSNR (dB)
0.75
16
0.7 0.65 0.6
14
TR1TPT-Econ t-SVD-TNN R1TP-Econ FP-LRCT
0.1
0.2
0.5 0.45 0.4
AN US
12
10
0.55
0.3
observation percentage
0.1
0.2
0.3
observation percentage
Figure 7: PSNR and SSIM versus observation percentage in the presence of noise.
completion under impulsive noise comparing to the noise-free case under the
M
similar parameter setting. In the remaining experiments, we set R = 15. Then the ‘t-SVD-TNN’ and ‘IRLS-t-SVD’ methods are included for comparison and
ED
for the IRLS based methods, p = 1.7 is used. The maximum number of iterations for the ‘t-SVD-TNN’ and ‘IRLS-t-SVD’ methods are 100 and 50. Furthermore, for ‘IRLS-t-SVD’ approach, the cases of tubal rank varying from 5 to 15 are
PT
tested and only the best result is used. The average PSNR and SSIM results of all the images are shown in Figure 10, and the average computational times of the ‘TIRLS-TR1TPT-Econ’, ‘t-SVD-TNN’ and ‘IRLS-t-SVD’ are 49.56s, 40.93s
CE
and 543.95s, respectively. It is seen that the ‘TIRLS-R1TPT-Econ’ approach can give the best recovery performance for both q2 = 10% and 20%, and the
AC
computational complexity is much smaller than the ‘IRLS-t-SVD’ approach, indicating the efficiency of the proposed method against this severe environment. The ‘t-SVD-TNN’ approach, on the other hand, fails to recover the images correctly. Finally, the video inpainting problem is tackled and the algorithms are evaluated on the widely used YUV Video Sequences [46]. In the experiment, we test 24
ACCEPTED MANUSCRIPT
TR1TPT-Econ
t-SVD-TNN
R1TP-Econ
FP-LRCT
Landscape
AN US
Rocks
Dock
CR IP T
Elephant
Doll
original image observed image
M
Figure 8: Typical recovery result of 5 images.
ED
the methods using three videos, namely, ‘Akiyo’, ‘Container’ and ‘Hall’, and the dimensions of the videos are 144×176×3×30. As the data are of 4-D, in order to apply the t-SVD based and proposed methods, we simply fold the 4th dimension
PT
into the second one. Furthermore, similar to the image cases, the 4-D video is reshaped to a higher-order tensor of dimensions 12 × 12 × 11 × 16 × 3 × 5 × 6, and
CE
then unfold the 3rd and 6th dimensions into the 1st one, and the 4th and 7th dimensions into the 2nd one to form a new tensor of dimensions 660 × 1152 × 3, and the final results of all the algorithms are the average of recovered values
AC
of both tensors. The parameters are the same as those in previous one except q1 = 50% and q2 = 10%. The average PSNR and SSIM results of all the 3 videos are given in Table 2, and one typical recovered frame of the ‘TIRLSTR1TPT-Econ’, ‘t-SVD-TNN’ and ‘IRLS-t-SVD’ approaches are displayed in Figure 11 at p = 1.5. We observe that the ‘TIRLS-TR1TPT-Econ’ scheme also
25
ACCEPTED MANUSCRIPT
Table 1: Completion results of different methods (RMSE (100 times mean ± std), -- means fail to give reasonable result). Uniform noise as outliers
CR IP T
SOR(dB) Algorithm
10
5
0
-5
-10
TIRLS-TR1TPT-Econ
4.58±0.28
5.72±0.54
7.07±1.11
8.22±1.47
10.31± 2.21
IRLS-t-SVD
11.71±1.51
13.36±1.73
15.70±1.78
16.84±2.01
18.37±2.21
IRLS-PARAFAC
98.57±1.22
98.67±1.17
98.44±1.48
98.93±1.09
98.78±1.21
t-SVD-TNN
17.61±0.27
32.03±0.45
56.52±0.79
--
--
Gaussian noise as outliers
10
5
0
-5
-10
TIRLS-TR1TPT-Econ
3.53± 0.20
4.34± 0.29
5.40± 0.44
6.82± 1.34
8.14± 1.94
IRLS-t-SVD
10.53±1.83
11.85±1.69
13.07± 1.89
14.96± 1.61
16.93±1.97
IRLS-PARAFAC
98.47±1.41
98.24±1.54
98.38±1.40
98.84±0.98
98.76±1.17
t-SVD-TNN
8.39±0.25
15.56±0.26
28.43±0.44
50.58±0.65
--
M
Algorithm
AN US
SOR(dB)
has good recovery performance and low complexity for a wide range of p, while
ED
the ‘IRLS-t-SVD’ method fails at p = 1.1. Generally speaking, the proposed ‘TIRLS-TR1TPT-Econ’ approach gives the best recovery and is the most computationally attractive one among all the algorithms, confirming its robustness
PT
and high efficiency against random tubal outliers.
CE
6. Conclusion
In this paper, we devise several data recovery methods for the tensor comple-
AC
tion problem using the idea of tensor tubal rank and greedy orthogonal tensor pursuit. The TR1TP and TR1TPT methods are first developed under the situation when the measured entries are randomly sampled or tubal randomly sampled. Using the idea of unraveling the multi-dimensional structure by performing FFT along the tensor tubes, the recovered orthogonal multi-dimensional tensors can be retrieved. Additionally, the TR1TP approach is proved to be con-
26
0.94
21.5
0.92
21
0.9
20.5
0.88
20
0.86
19.5 19
CR IP T
22
SSIM
PSNR (dB)
ACCEPTED MANUSCRIPT
0.84 0.82
TIRLS-TR1TPT-Econ (q1 = 60%)
18.5
0.8
TIRLS-TR1TPT-Econ (q1 = 40%)
18
0
5
10
15
iteration number
0.76
AN US
17.5
0.78
20
0
5
10
15
20
iteration number
Figure 9: PSNR and SSIM versus number of iterations.
verged linearly. Furthermore, when there are tubal outliers in the measurements,
M
the algorithm is modified by combining the idea of tubal `p -norm optimization. Experimental results show that the proposed methods outperform the state-
complexity.
ED
of-the art algorithms in terms of RMSE, PSNR, SSIM and/or computational
PT
7. Appendix A
We show that the proposed TR1PT algorithm has a linear convergence rate.
CE
Before proving the convergence, some useful properties will be presented first. Mi )Ω >= 0 and < R (k+1) , M i >= 0 for i = Property A.1. < R (k+1) , (M
AC
1, 2, · · · , k.
Proof. Note that θ (k) is the optimal solution of problem (25). As a result, we
M(i) (θ (i) ))Ω , (M Mi )Ω >= 0 for i = 1, 2, · · · , k, where (M M(i) (θ (i) ))Ω have < Y Ω −(M
X (k) )Ω and X (k) = M (k) (θ k ), is defined in (15). As long as R (k+1) = Y Ω − (X
Mi )Ω >= 0. Furthermore, the we reach to the conclusion that < R (k+1) , (M
unobserved entries in R(k+1) are all set to 0 according to the definition, leading 27
ACCEPTED MANUSCRIPT
22
0.95
21
0.9
20 0.85
CR IP T
0.8
18
SSIM
PSNR (dB)
19
17
TIRLS-TR1TPT-Econ (q =10%) 2
0.75
t-SVD-TNN (q2 =10%)
IRLS-t-SVD (q2 =10%)
16
TIRLS-TR1TPT-Econ (q2 =20%)
0.7
t-SVD-TNN (q2 =20%)
15
IRLS-t-SVD (q2 =20%)
0.65
13 0.2
0.3
0.4
0.5
observation percentage
AN US
14
0.6 0.2
0.6
0.3
0.4
0.5
0.6
observation percentage
Figure 10: PSNR and SSIM versus observation percentage under salt-and-pepper noise.
Mi ) >= 0 for i = 1, 2, · · · , k. to < R (k+1) , (M
M
This property indicates that R (k+1) is perpendicular to the previously re-
covered M i .
ED
˙ (k) in (25) has a full column rank. Property A.2. For R (k) 6= 0, M
Proof. According to the definition, when R (k) 6= 0, we have R (i) 6= 0 for all i = 1, 2, · · · , k − 1, otherwise the stopping criterion of the proposed method
PT
˙ (1) 6= 0. Hence, the conclusion is violated. Since R (1) = Y Ω 6= 0, we have M ˙ (k−1) , k ≥ 2, holds for k = 1. We now assume that the property holds for M
CE
˙ (k) . In particular, as M ˙ (k−1) has full column and show that it also holds for M
AC
˙ (k) does not have full column rank, then rank, suppose that for contradiction, M h iT ˙ (k−1)α or ˙k=M there exists an α = α1 α2 · · · αk−1 that satisfies m Mk )Ω = (M
i=k−1 X i=1
Mi )Ω . αi (M
According to Property A.1., we have < R (k+1) , M k
R(k)
R
, M k >= 0, that is, <
R (k) , M
k
(40) >= 0 and <
>= 0 in (22), indicating that the maximum
singular value of all the frontal slices of R(k) is 0, leading to R(k) = R(k) = 0. 28
ACCEPTED MANUSCRIPT
Table 2: PSNR(dB), SSIM and computational complexity (seconds) of different methods under different p. p
Algorithm
1.1
1.3
1.5
1.7
TIRLS-TR1TPT-Econ
20.44
21.86
22.05
IRLS-t-SVD
11.82
17.92
19.98
t-SVD-TNN
16.77
SSIM
p 1.1
1.3
1.5
TIRLS-TR1TPT-Econ
0.900
0.923
0.927
IRLS-t-SVD
0.505
0.815
0.875
t-SVD-TNN
20.21
19.83
19.50
18.67
1.7
1.9
0.899
0.875
0.863
0.838
AN US
Algorithm
1.9
CR IP T
PSNR(dB)
0.771
Complexity(seconds)
p
1.1
1.3
1.5
1.7
1.9
TIRLS-TR1TPT-Econ
129.2
129.8
129.5
127.6
129.6
IRLS-t-SVD
950.8
1358.9
1359.3
1391.1
1245.3
t-SVD-TNN
M
Algorithm
395.9
ED
˙ (k) has a full column This contradicts the fact that R (k) 6= 0. Therefore, M rank.
R(k+1) || and Now we can build the relation between two successive residuals ||R
PT
(k) R(k) || by defining θk−1 ||R , namely, the value of θ(k) at the (k − 1)-th iteration,
CE
as 0. Setting θ (k) = θ (k−1) + ι(k) , it follows from (18) that R(k) − ι(k) = arg min ||R
ι
i=1
(41)
AC
ι
k X ˙ (k) ι(k) ||2 Mi )Ω ||22 = arg min ||r(k) − M {ι(k) }(i) (M 2
˙ (k) is defined in (25). Since M ˙ (k) has full R(k) )} and M where r(k) = vec{(R
column rank according to Property A.2, we have the solution ˙ (k) )T M ˙ (k) )−1 (M ˙ (k) )T r(k) . ι(k) = ((M
29
(42)
ACCEPTED MANUSCRIPT
Observed frame
TIRLS-TR1TPT-Econ
IRLS-t-SVD
t-SVD-TNN
AN US
Hall
CR IP T
Container
Akiyo
Original frame
Figure 11: Typical recovered frame of 3 videos.
Defining
M(k) (ι(k) ))Ω = L (k) = (M
k X i=1
Mi )Ω {ι(k) }(i) (M
(43)
M
we have X k = X k−1 + L (k) and R (k+1) = R (k) − L (k) .
R(k+1) ||22 = ||R R(k) ||22 − ||L L((k)) ||22 and ||L L(k) ||22 ≥< ||R
R (k) , M k >.
ED
Property A.3.
According to Property A.1 and (43), it is clear that < R (k+1) , L (k) >= 0.
Hence, we obtain
PT
R(k+1) ||22 = ||R R(k) − L (k) ||22 ||R
AC
CE
R(k) ||22 − 2 < R (k) , L (k) > +||L L(k) ||22 = ||R R(k) ||22 − 2 < R (k+1) + L (k) , L (k) > +||L L(k) ||22 = ||R R(k) ||22 − ||L L(k) ||22 . = ||R
(44)
L(k) ||22 is bounded as follows. For any R (k) 6= 0, according to Property The ||L
A.1, we have h ˙ (k) )T r(k) = 0 (M
0
···
0
< R (k) , M k >
iT
.
(45)
From the definition, M i for i = 1, 2, · · · , k is normalized to 1. As long as the
˙ (k) , denoted as m Mk )Ω }, contains only part of ˙ k = vec{(M k-th column of M 30
ACCEPTED MANUSCRIPT
˙ (k) = QU be the QR ˙ k || ≤ 1. Let M elements of M k , it has the property ||m
˙ k where QT Q = IM , MΩ = M1 M2 M3 q1 , and U ∈ Rk×k is factorization of M Ω an upper triangular matrix. According to [33], the (k, k) element of U, denoted
CR IP T
as Ukk , has the property ˙ k || ≤ 1. |Ukk | ≤ ||m
(46)
Combining (42) and (46), we have
˙ (k) )T r(k) )k =< R (k) , M k > /|Ukk |2 ≥< R (k) , M k > . {ι(k) }(k) = ((UT U)−1 (M
(47)
L(k) ||22 =< R (k) − R (k+1) , ||L
AN US
L(k) ||22 can be written as Furthermore, ||L
k X Mi )Ω > {ι(k) }(i) (M i=1
Mk )Ω > + < R (k) , =< R (k) , ιk,k (M (k)
}(k) < R
≥< R (k) , M k > .
, Mk >
i=1
Mi )Ω > − < R (k+1) , {ι(k) }(i) (M
M
= {ι
(k)
k−1 X
k X Mi )Ω > {ι(k) }(i) (M i=1
(48)
ED
Now the convergence of the proposed algorithm is investigated. According to the definition, we have
M3 {m3 } {m } 1 1 X g (k) gk 3 , R < R (k) , M k >= M3 M3 1
PT
< R (k) , M k > =
CE
=
M3 {m } {m3 } p {m3 } 1 X g H ){m3 } , R fk 3 > (Sfk M3 /||Sfk ||F ) < Ufk (V k M3 1
(49)
{m }
AC
3 {m3 } {m3 } g (k) where Sfk is the dominant singular value of R defined in (23), Ufk g H ){m3 } are the corresponding singular vectors, and ||S fk || is defined in and (V k
(24). It follows that Sfk
{m3 }
{m3 }
{m } g H ){m3 } , R fk 3 >≥ (V k
= < Ufk r {m } fk 3 ||2 /min{M1 , M2 } ≥ ||R 2
31
r
fk ||R
{m3 }
fk ||22 /rank(R
{m3 }
) (50)
ACCEPTED MANUSCRIPT
and ||Sfk ||22 =
M3 X
(Sfk
{m3 }
m3 =1
)2 ≥
{m }
fk 3 ||2 fk ||2 ||R ||R 2 2 = . min{M , M } min{M , M2 } 1 2 1 =1
M3 X
m3
(51)
CR IP T
Combining the equations (49)-(51) yields g (k) ||2 /(M min{M , M }) = ||R R(k) ||22 /min{M1 , M2 } (52) < R (k) , M k >≥ ||R 3 1 2 2
which in turn leads to
R(k+1) ||22 = ||R R(k) ||22 − ||L L(k) ||22 ≤ ||R R(k) ||22 − < R (k) , M k > ||R Rk ||22 . ≤ (1 − 1/ min{M1 , M2 })||R
(53)
AN US
Hence, the TR1TP algorithm converges linearly and is not dependent on
the length of the third dimension M3 . Furthermore, by using the fact that R(1) ||2 = ||Y Y Ω ||2 , we get (26). ||R References
M
References
[1] N. Boumal and P. A. Absil, “RTRMC: A Riemannian trust-region method
ED
for low-rank matrix completion,” Proceedings of International Conference on Neural Information Processing Systems 24 (NIPS 2011), pp.406-414, Dec. 2011, Granada, Spain.
PT
[2] N. Linial, E. London and Y. Rabinovich, “The geometry of graphs and some of its algorithmic applications,” Combinatorica, vol.15, no.2, pp.215-
CE
245, Jun. 1995.
[3] Y. Pang, S. Wang and Y. Yuan, “Learning regularized LDA by cluster-
AC
ing,” IEEE Transactions on Neural Networks and Learning Systems, vol.25, no.12, pp.2191-2201, Apr. 2014.
[4] M.J. Taghizadeh, R. Parhizkar, P.N. Garner, H. Bourlard and A. Asaei, “Ad hoc microphone array calibration: Euclidean distance matrix completion algorithm and theoretical guarantees,” Signal Processing, vol.107, pp.123-140, Feb. 2015. 32
ACCEPTED MANUSCRIPT
[5] L.T. Huang, A.L.F.De Almeida and H.C. So, “Target estimation in bistatic MIMO radar via tensor completion,” Signal Processing, vol.120, pp.654659, Mar. 2016.
CR IP T
[6] X. Hu, N. Tong, J. Wang, S. Ding and X. Zhao, “Matrix completion-based MIMO radar imaging with sparse planar array,” Signal Processing, vol.131, pp.49-57, Feb. 2017.
[7] T. G. Kolda and J. Sun, “Scalable tensor decompositions for multi-aspect data mining,” 2008 Eighth IEEE International Conference on Data Mining,
AN US
pp.363-372, Dec. 2008, Pisa, Italy.
[8] J. Sun, S. Papadimitriou, C.-Y. Lin, N. Cao, S. Liu and W. Qian, “MultiVis: Content-based social network exploration through multi-way visual analysis,” SIAM International Conference on Data Mining (SDM2009), pp.1063-1074, Sparks, Nevada, USA, Apr. - May, 2009.
M
[9] G. Obozinski, B. Taskar and M. I. Jordan, “Joint covariate selection and joint subspace selection for multiple classification problems,” Statistics
ED
Computing, vol.20, no.2, pp.231-252, 2010. [10] J. Liu, P. Musialski, P. Wonka and J. Ye, “Tensor completion for estimating missing values in visual data,” IEEE Transactions on Pattern Analysis and
PT
Machine Intelligence, vol.35, no.1, pp.208-220, 2013. [11] Z. Zhang, G. Ely, S. Aeron, N. Hao and M. Kilmer, “Novel methods for mul-
CE
tilinear data completion and de-noising based on tensor-SVD,” IEEE Conference on Computer Vision and Pattern Recognition (CVPR)), pp.3842-
AC
3849, Jun. 2014.
[12] M. Fazel, H. Hindi and S. Boyd, “A rank minimization heuristic with application to minimum order system approximation,” Proceedings of the 2001 American Control Conference, vol.6, no.2, pp. 4734-4739, Arlington, USA, 25-27 Jun. 2001.
33
ACCEPTED MANUSCRIPT
[13] B. Recht, M. Fazel and P.A. Parrilo, “Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization,” SIAM Review, vol.52, pp.471-501, 2010.
CR IP T
[14] E.J. Cand` es and B. Recht, “Exact matrix completion via convex optimiza-
tion,” Foundations of Computational Mathematics, vol.9, no.6, pp.717, 2009.
[15] J.F. Cai, E.J. Cand` es and Z. Shen, “A singular value thresholding algo-
rithm for matrix completion,” SIAM Journal on Optimization, vol.20, no.4,
AN US
pp.1956-1982, Jan. 2010.
[16] R.H. Keshavan and S. Oh, “A gradient descent algorithm on the Grassman manifold for matrix completion,” arXiv preprint arXiv:0910.5260, 2009. [17] P. Jain, R. Meka and I.S. Dhillon, “Guaranteed rank minimization via singular value projection,” Proceedings of International Conference on Neural
Vancouver, Canada.
M
Information Processing Systems 23 (NIPS 2010), pp. 937-945, Dec. 2010,
ED
[18] S.Q. Ma, D. Goldfarb and L. F. Chen, “Fixed point and Bregman iterative methods for matrix rank minimization,” Mathematical Programming,
PT
vol.128, pp.321-353, 2011.
[19] T.G. Kolda and B.W. Bader, “Tensor decompositions and applications,”
CE
SIAM Review, vol.51, no.3, pp.455-500, Sep. 2009. [20] C. Mu, B. Huang, J. Wright and D. Goldfarb, “Square deal: Lower bounds
AC
and improved relaxations for tensor recovery,” Proceedings of the 31st International Conference on Machine Learning (ICML), wol.32, no.2, pp.7381, 22-24 Jun. 2014, Beijin, China.
[21] P. Jain and S. Oh, “Provable tensor factorization with missing data,” Proceedings of International Conference on Neural Information Processing Systems 27 (NIPS 2014), pp.1431-1439, Dec. 2014, Montr´ eal, Canada.
34
ACCEPTED MANUSCRIPT
[22] V. De Silva and L.H. Lim, “Tensor rank and the ill-posedness of the best low-rank approximation problem,” SIAM Journal on Matrix Analysis Applications, vol.30, no.3, pp.1084-1127, 2008.
CR IP T
[23] M. Filipovi´ c and A. Juki´ c, “Tucker factorization with missing data with application to low-n-rank tensor completion,” Multidimensional Systems and Signal Processing, vol.26, no.2, pp.677-692, Jul. 2015.
[24] T.N. Thieu, H.J. Yang, T.D. Vu and S.H. Kim, “Recovering incomplete data using Tucker model for tensor with low-n-rank,” International Journal
AN US
of Contents, vol.12, no.23 pp.22-28, Sep. 2016.
[25] L. Yang, J. Fang, H. Li and B. Zeng, “An iterative reweighted method for Tucker decomposition of incomplete tensors,” IEEE Transactions on Signal Processing, vol.64, no.18, pp.4817-4829, Sep. 2016.
[26] L. Yang, Z.H. Huang and X. Shi, “A fixed point iterative method for low-
M
n-rank tensor pursuit,” IEEE Transactions on Signal Processing, vol.61, no.11, pp.2952-2962, Jun. 2013.
ED
[27] Y. Ming and C. H. Zhang, “On tensor completion via nuclear norm minimization,” Foundations of Computational Mathematics, vol.16, no.4,
PT
pp.1031-1068, Aug. 2016.
[28] Y. Yang, Y. Feng and J. A. K. Suykens, “A rank-one tensor updating algorithm for tensor completion,” IEEE Signal Processing Letters, vol.22,
CE
no.10, pp.1633-1637, Oct. 2015.
AC
[29] M. Kilmer, K. Braman, N. Hao and R. Hoover, “Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging,” SIAM Journal on Matrix Analysis and Applications, vol.34, no.1, pp.148-172, Jan. 2013.
[30] Z. Zhang and S. Aeron, “Exact tensor completion using t-SVD,” IEEE Transactions on Signal Processing, vol.65, no.6, pp.1511-1526, Dec. 2016.
35
ACCEPTED MANUSCRIPT
[31] M. Yuan and Y. Lin, “Model selection and estimation in regression with grouped variables,” Journal of the Royal Statistical Society, vol.68, no.1, pp.49-67, 2006.
CR IP T
[32] W. Deng, W. Yin and Y. Zhang, “Group sparse optimization by alternating direction method,” Proceedings of SPIE, Wavelets and Sparsity XV, vol.8858, 26 Sep. 2013.
[33] Z. Wang, M.J. Lai, Z. Lu, W. Fan, H. Davulcu and J.P. Ye, “Orthogonal
rank-one matrix pursuit for low rank matrix completion,” SIAM Journal
AN US
on Scientific Computing, vol.37, no.1, 2014.
[34] X. Liao, H. Li and L. Carin, “Generalized alternating projection for weighted-`2,1 minimization with applications to model-based compressive sensing,” SIAM Journal on Imaging Sciences, vol.7, no. 2, pp.797-823, 2014. [35] C.D. Martin, R. Shafer, and B. LaRue, “An order-p tensor factorization
M
with applications in imaging,” SIAM Journal on Scientific Computing, vol.35, no.1, pp.474-490, 2013.
ED
[36] M. Jaggi and M. Sulovsk, “A simple algorithm for nuclear norm regularized problems,” International Conference on Machine Learning (ICML), pp.471478, Jul. 2010, Haifa, Israel.
PT
[37] X. Zhang, Y. Yu and D. Schuurmans, “Accelerated training for matrixnorm regularization: A boosting approach,” Proceedings of International
CE
Conference on Neural Information Processing Systems 25 (NIPS 2012), pp.2915-2923, 3-6 Dec. 2012, Lake Tahoe, Nevada, USA.
AC
[38] X.Y. Liu, S. Aeron, V. Aggarwal, X. Wang and M.Y. Wu, “Fine-grained indoor localization with adaptively sampled RF fingerprints,” Hungarian Journal of Industry and Chemistry, vol.58, no.3, 2015.
[39] Y. Pang, X. Li and Y. Yuan, “Robust tensor analysis with L1-norm,” IEEE Transactions on Circuits and Systems for Video Technology, vol.20, no.2, pp.172-178, Feb. 2010. 36
ACCEPTED MANUSCRIPT
[40] Y. Yang, S. Mehrkanoon and J.A.K. Suykens, “Higher order matching pursuit for low rank tensor learning,” arXiv:1503.02216 [stat.ML], Mar. 2015. [41] Z. Wang, A.C. Bovik, H.R. Sheikh, and E.P. Simoncelli, “Image quality as-
CR IP T
sessment: From error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600-612, Apr. 2004.
[42] D. Martin, C. Fowlkes, D. Tal, and J. Malik, “A database of human seg-
mented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in Proceedings of IEEE In-
AN US
ternational Conference of Computer Vision (ICCV), vol. 2, no. 11, pp. 416-423, 2001, Barcelona, Spain.
[43] W. Sun, Y. Chen, L. Huang and H.C. So, “Tensor completion via generalized tensor tubal rank minimization using general unfolding,” IEEE Signal Processing Letters, Vol.25, No.6, pp.868-872, Jun. 2018.
M
[44] W. Sun, X. Lin, H.C. So and L. Huang, “Iteratively reweighted tensor SVD for robust multi-dimensional harmonic retrieval,” Proceedings of the Inter-
ED
national Conference on Acoustics, Speech, and Signal Processing (ICASSP 2016), pp. 4312-3512, Mar. 2016, Shanghai, China. [45] X. Lin, L. Huang and W. Sun, “Lp-PARAFAC for joint DoD and DoA
PT
estimation in bistatic MIMO radar,” Proceedings of 2016 CIE International Conference on Radar, pp.1-5, Oct. 2016, Guangzhou, China.
AC
CE
[46] http://trace.eas.asu.edu/yuv/.
37