Preprint
Article

Tensor Conjugate-Gradient Methods With Automatically Determination of Regularization Parameters for Ill-Posed Problems With T-product

Altmetrics

Downloads

105

Views

28

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

26 October 2023

Posted:

26 October 2023

You are already at the latest version

Alerts
Abstract
This paper presents three types of tensor Conjugate-Gradient methods for solving large-scale linear discrete ill-posed problems based on the t-product between third-order tensors. An automatical determination strategy of a suitable regularization parameter is proposed for the tensor conjugate gradient (tCG) method. A truncated version and a preprocessed verion of the tCG method are further presented. The discrepancy principle is employed to determine a suitable regularization parameter. Several numerical examples are given to show the effectiveness of the proposed tCG methods in image and video restoration.
Keywords: 
Subject: Computer Science and Mathematics  -   Computational Mathematics

1. Introduction

Tensors are high-dimensional arrays that have many applications in science and engineering, including in image, video and signal processing, computer vision, and network analysis [11,12,16,17,18,19,20,26]. A new t-product based on third-order tensors proposed by Kilmer et al [1,2]. When using high-dimensional data, t-product shows a greater potential value than matricization, see [2,6,11,12,21,22,24,25,27]. The t-product has been found to have special value in many application fields, including image deblurring problems [1,6,11,12], image and video compression [26], facial recognition problems [2], etc.
In this paper, we consider the solution of large minimization problems of the form
min X R m × 1 × n A X B F , A = [ a ] i , j , k = 1 l , m , n R l × m × n , B R l × 1 × n .
The Frobenius norm of singular tube of A rapidly attenuates to zero with the increase of the index number. In particular, A has ill-determined tubal rank. Many of its singular tubes are nonvanishing with tiny Frobenius norm of different orders of magnitude. Problems (1) with such a tensor is called the tensor discrete linear ill-posed problems. They arise from the restoration of color image and video, see e.g., [1,11,12]. Throughout this paper, the operation ∗ represents tensor t-product and · F denotes the tensor Frobenius norm or the spectral matrix norm.
We assume that the observed tensor B R m × 1 × n is polluted by an error tensor E R m × 1 × n , i.e.,
B = B t r u e + E ,
where B t r u e R m × 1 × n is an unknown and unavailable error-free tensor related to B . B t r u e is determined by A X = B t r u e , where X t r u e represents the explicit solution of problems (1) that is to be found. We assume that the upper bound of the Frobenius norm of E is known, i.e,  
E F δ .
Straightforward solution of (1) is usually meanless to get an approximation of B t r u e because of the illposeness of A = [ a ] i , j , k = 1 l , m , n and the error E will be amplified severely. We use Tikhonov regularization to reduce this effect in this paper and replace (1) with penalty least-squares problems
min X R m × 1 × n A X B F 2 + μ X F 2 ,
where μ is a regularization parameter. We assume that
N ( A ) N ( I ) = O ,
where N ( A ) denotes the null space of A , I is the identity tensor and O R m × 1 × n is a lateral slice whose elements are all zero. The normal equation of minimization problem (4) is
( A T A + μ I ) X = A T B ,
then
X μ = A T A + μ I 1 A T B
is the unique solution of the Tikhonov minimization problem (4) under the assumption (5).
There are many techniques to determine the regularization parameter μ , such as the L-curve criterion, generalized cross validation (GCV), and the discrepancy principle. We refer to [4,5,8,9,10] for more details. In this paper, the discrepancy principle is extended to tensors based on t-product and is employed to determine a suitable μ in (4). The solution X μ of (4) satisfies
A X μ B F η δ ,
where η > 1 is usually a user-specified constant and is independent of δ in (3). When E F is smaller enough, and δ approaches 0, result in X μ X t r u e . For more details on the discrepancy principle, see e.g., [7].
In this paper, we also consider the expansion of minimization problem (1) of the form
min X R m × p × n A X B F 2 + μ X F 2 ,
where B R m × p × n , p > 1 .
There are many methods for solving large-scale discrete linear ill-posed problems (1). Recently, a tensor Golub- Kahan bidiagonalization method [11] and a GMRES method [12] were introduced for solving large-scale linear ill-posed problems (4). The randomized tensor singular value decomposition (rt-SVD) method in [3] was presented for computing super large data sets, and has prospects in image data compression and analysis. Ugwu and Reichel [23] proposed a new random tensor singular value decomposition (R-tSVD), which improves the truncated tensor singular value decomposition (T-tSVD) in [1]. Kilmer et al. [2] presented a tensor Conjugate-Gradient (t-CG) method for tensor linear systems A X = B corresponding to the least-squares problems. The regularization parameter in the t-CG method is user-specified. In this paper, we further discuss the automatical determinization of suitable regularization parameters of the tCG method by the discrepancy principle. The proposed method is called the tCG method with automatical determination of regularization parameters (auto-tCG). We also present a truncated auto-tCG method (auto-ttCG) to improve the auto-tCG method by reducing the computation. At last, a preprocessed version of the auto-ttCG method is proposed, which is abbreviated as auto-ttpCG.
The rest of this paper is organized as follows. Section 2 introduces some symbols and preliminary knowledge that will be used in the context. Section 3 presents the auto-tCG, auto-ttCG and auto-ttpCG methods for solving the minimization problems (4) and (9). Section 4 gives several examples on image and video restoration and Section 5 draws some conclusions.

2. Preliminaries

This section gives some notations and definitions, and briefly summarizes some results that will be used later. For a third-order tensor A R l × m × n , Figure 1 shows the frontal slices A ( : , : , k ) , lateral slices A ( : , j , : ) and tube fibers A ( i , j , : ) . We abbreviate A k = A ( : , : , k ) for simplication. An l n × m matrix is obtained by the operator unfold ( A ) , whereas the operator fold folds this matrix back to the tensor A , i.e.,
unfold A = A 1 A 2 A n , fold unfold A = A .
Definition 1. 
Let A R l × m × n , then a block-circulant matrix of A is denoted by bcirc ( A ) , i.e.,
bcirc A = A 1 A 2 A n A n A 1 A n 1 A 2 A 3 A 1 .
Definition 2. ([1]) Given two tensors A R l × m × n and B R m × p × n , the t-product A B is defined as
A B = fold ( bcirc ( A ) unfold ( B ) ) = C ,
where C R l × p × n .
The following remarks will be used in Section 3.
Remark 1. ([14]) For suitable tensors A and B , it holds that
(1). bcirc ( A B ) = bcirc ( A ) bcirc ( B ) .
(2). bcirc ( A T ) = bcirc ( A ) T .
(3). bcirc ( A + B ) = bcirc ( A ) + bcirc ( B ) .
Let F n be an n-by-n unitary discrete Fourier transform matrix, i.e,  
F n = 1 n 1 1 1 1 1 ω ω 2 ω n 1 1 ω 2 ω 4 ω 2 n 1 1 ω n 1 ω 2 ( n 1 ) ω n 1 n 1 ,
where ω = e 2 π i n , then we get the tensor A ^ generated by using FFT along each tube of A , i.e,
bdiag A ^ = A ^ 1 A ^ 2 A ^ n = F n I l bcirc A F n * I m ,
where ⊗ is the Kronecker product, F n * is the conjugate transposition of F n and A ^ i denetes the frontal slices of A ^ . Thus the t-product of A and B in (10) can be expressed by
A B = fold F n * I l F n I l bcirc A F n * I m F n I m unfold B ,
and (10) is reformulated as
A ^ 1 A ^ 2 A ^ n B ^ 1 B ^ 2 B ^ n = C ^ 1 C ^ 2 C ^ n .
It is easy to calculate (12) in MATLAB.
For a non-zero tensor X R m × 1 × n , we can decompose it in the form
X = D d ,
where D R m × 1 × n is a normalized tensor; see, e.g., [6] and d R 1 × 1 × n is a tube scalar. Algorithm 1 summarizes the decomposition in (14).
Algorithm 1 Normalization
Input: X R m × 1 × n is a nonzero tensor
Output: D , d with X = D d , D = 1
D fft( X ,[ ],3)
for  j = 1 , 2 , , n   do
     d j D j 2 ( D j is a vector)
    if  d j > t o l  then
         D j 1 d j D j
    else
         D j randn ( m , 1 ) ; d j D j 2 ; D j 1 d j D j ; d j 0
    end if
end for
D ifft( D ,[ ],3); d ifft( d ,[ ],3)
Given a tensor A R l × m × n , the singular value decomposition (tSVD) of A is expressed as
A = U S V T ,
where U R l × l × n and V R m × m × n are orthogonal under t-product,
S = diag [ s 1 , s 2 , . . . , s min { l , m } ] R m × l × n
is an upper triangular tensor with the singular tubes s j satisfying
s 1 F s 2 F s min l , m F .
The operators squeeze and twist [13] are expressed by
X = squeeze ( X j ) X ( i , j ) = X ( i , 1 , j ) , twist ( squeeze ( X ) ) = X .
Figure 2 illustrates the transformation between a matrix and a tensor column by using squeeze and twist . Generally, the operators multi squeeze and multi twist are defined for a third-order tensor to make it squeezed or twisted. For a tensor D R m × p × n with p > 1 , C = multi squeeze ( D ) means that all side slices of D are squeezed and stacked as front slices of C , the operator multi twist is the reverse operation of multi squeeze . Thus multi twist ( multi squeeze ( D ) ) = D . We refer to Table 1 for more notations and definitions.

3. Tensor Conjugate-Gradient methods

This section first discusses the automatical determination of a suitable regularization parameter for the tensor conjugate gradient (tCG) method presented by Kilmer et al. in [13]. We abbreviate the improved method as auto-tCG. A truncated auto-tCG method is developed to improve the auto-tCG method and is abbreviated as auto-ttCG. A preprocessed version of the auto-ttCG method is presented, which is abbreviated as auto-ttpCG.

3.1. The auto-tCG Method

The tensor Conjugate-Gradient (t-CG) method is presented in [2] for the least-squares solution of the tensor linear systems A X = B . The regularization parameter in the t-CG method was not discussed and was user-specified. This subsection improves the t-CG method by employing the discrepancy principle to determine a suitable regularization parameter under the assumption (3) and uses it to solve the normal equation (6). We consider the polynomial function
μ k = μ 0 q k , k = 0 , 1 , ,
where q ( 0 , 1 ) . We set μ 0 = A F , and obtain an optimal regularization parameter by continuously reducing the parameter. An effective method to deal with the general problems (9) is to regard it as p independent subproblems (4), i.e.,
min X j R m × 1 × n A X j B j F 2 + μ X j F 2 , j = 1 , , p ,
where B j is the tensor column of the tensor B and is polluted by the noise E j . B j , t r u e represents unknown error-free tensor. Assume the noise tensor
E j = B j B j , t r u e
can be used or the norm of E j can be estimated, i.e.,
E j F δ j , j = 1 , , p .
Algorithm 2 summarizes the auto-tCG method for solving (9). The initial tensor of Algorithm 2 is set as zero tensor. The iteration is stopped when the Frobenius norm of the residual tensor
R j , μ k i = A T B j ( A T A + μ k I ) X j , μ k i
is small enough, where R j , μ k i denotes the residual generated by the i-th iterative solution X j , μ k i of the normal equation with μ k of the j-th independent subproblem. Let X i n t = X μ k * be the initial tensor of the normal equation of μ k + 1 . When μ = μ k with m iterations for the CG-process, the affine space is X μ k 0 + K m A T A + μ k I , r μ k 0 , where r μ k 0 = A T B A T A + μ k I X μ k 0 .
Algorithm 2 The auto-tCG method for sloving (9).
Input: A R m × m × n , B j R m × 1 × n , δ j , j = 1 , . . . , p , μ 0 , η > 1 .
Output: Approximate solution X * of problem (9).
for  j = 1 , 2 , . . . p   do
     X i n t = 0 , k = 0 .
    while  do A X j , μ k * B j F 2 > η 2 δ j 2
         k = k + 1 , ( A T A + μ k I ) X j = A T B j , e.g., μ k = μ 0 q k
         [ R 0 , a ] Normalize ( A T B j ( A T A + μ k I ) X i n t ) ; P 0 R 0 .
         i = 0 , σ > t o l .
        while  σ > t o l  do
            i = i + 1 .
            c = P i 1 T ( A T A + μ k I ) P i 1 1 R i 1 T R i 1 .
            X i = X i 1 + P i 1 c .
            R i = R i 1 ( A T A + μ k I ) P i + 1 c .
            σ = | R i F R i 1 F | .
            d = R i 1 T R i 1 1 R i T R i .
            P i = R i + P i 1 d .
        end while
         X j , μ k * = X i a ( X j , μ k * is the solution of the normal equation about μ k of the j-th independent subproblem (4)).
         X i n t = X j , μ k * .
    end while
     X ( : , j , : ) * = X j , μ k * .
end for

3.2. The truncated tensor Conjugate-Gradient method

Frommer and Maass in [15] proposed a good condition that can roughly judge some inappropriate value of μ . We introduce this condition to improve Algorithm 2 by excluding some unsuitable value of μ , and present a truncated tensor conjugate-gradient method for solving (9). We first give the following results.
Theorem 1. 
Given A R l × m × n , define a t-linear operator T: R m × 1 × n R l × 1 × n , i.e., T ( X ) = A X with X R m × 1 × n . Let X μ * be the exact solution of the normal equations
( A T A + μ I ) X = A T B ,
then for an arbitrary X R m × 1 × n , we have
A X μ * B F 2 A X B F 2 1 4 μ A T B ( A T A + μ I ) X F 2 .
Proof. 
For an arbitrary X R m × 1 × n , set Z = X μ * X . Let the singular value decomposition of A be A = U S V T , then
A Z = U S V T Z .
Suppose V T Z = D R m × 1 × n , then
A Z F 2 = U S V T Z F 2 = S D F 2 = bcirc ( S ) unfold ( D ) 2 2 .
Thus
( A T A + μ I ) Z F 2 = V ( S T S + μ I ) V T Z F 2 = V ( S T S + μ I ) D F 2 = S T S + μ I D F 2 = ( bcirc ( S T * S ) + μ bcirc ( I ) ) unfold ( D ) 2 2 = ( bcirc ( S ) T bcirc ( S ) + μ bcirc ( I ) ) unfold ( D ) 2 2 .
Denote bcirc ( S ) = S R n l × n m , bcirc ( I ) = I R n m × n m and unfold ( D ) = d R n m × 1 , then A Z F 2 = S d 2 2 and ( A T A + μ I ) Z F 2 = ( S T S + μ I ) d 2 2 . Thus we transform the tensor norm into the equivalent matrix norm. Let the singular value decomposition of S be S = U Σ V T , where Σ = diag σ 1 , σ 2 , . . . , σ r , r min n l , n m , U = [ u 1 , u 2 , . . . , u r ] and V = [ v 1 , v 2 , . . . , v r ] are orthogonal matrices with orthogonal columns u k R n l × 1 and v k R n m × 1 , respectively. Thus we have
S d = σ k > 0 σ k d , v k u k .
Using the equation s 2 = ( s + μ s 1 ) 2 ( s 2 + μ ) 2 with the estimate
1 s + μ s 1 1 2 μ , ( s , μ > 0 ) ,
we have
Sd 2 2 = σ k > 0 σ k 2 d , v k 2 = σ k > 0 σ k + μ σ k 1 2 σ k 2 + μ 2 d , v k 2 1 4 μ σ k > 0 σ k 2 + μ 2 d , v k 2 .
Note that
S T S + μ I d 2 2 = σ k > 0 σ k 2 + μ 2 d , v k 2 ,
It results from (19) and (20) that
Sd 2 2 1 4 μ S T S + μ I d 2 2 .
Note that A Z F 2 = S d 2 2 and ( A T A + μ I ) Z F 2 = ( S T S + μ I ) d 2 2 , we have
A Z F 2 1 4 μ ( A T A + μ I ) Z F 2 .
Thus
A X μ B F 2 = A X B + A X μ * X F 2 A X B F 2 A Z F 2 A X B F 2 1 4 μ A T A + μ I Z F 2
Note that
( A T A + μ I ) Z = ( A T A + μ I ) ( X μ * X ) = A T B ( A T A + μ I ) X ,
then (23) and (22) result in
A X μ * B F 2 A X B F 2 1 4 μ A T B ( A T A + μ I ) X F 2 .
   □
We will apply Theorem 1 to predict in advance whether the exact solution X μ k * satisfies the discrepancy principle in Algorithm 2. We add the condition
A X μ k i B F 2 1 4 μ k R μ k i F 2 > η 2 δ 2
in steps 9-16 of Algorithm 2. If the i-th iteration solution of the normal equation with μ k is X μ k i and its residual R μ k i satisfies (24), then A X μ k * B F 2 > η 2 δ 2 . This indicates that the exact solution of the normal equation with μ k does not satisfy the discrepancy principle, so continue to solve next normal equation with μ k + 1 . Therefore, we obtain a truncated tensor conjugate-gradient method of automatical determination of a suitable regularization parameter, which is abbreviated as auto-ttCG. Algorithm 3 summarizes the auto-ttCG method.
Algorithm 3 The auto-ttCG method for sloving (9)
Input: A R m × m × n , B j R m × 1 × n , δ j , j = 1 , . . . , p , μ 0 , η > 1 , tol.
Output: Approximate solution X * of problem (9).
for  j = 1 , 2 , . . . p   do
     X i n t = 0 , k = 0
    while  A X j , μ k i B j F 2 > η 2 δ j 2  do
         k = k + 1 , ( A T A + μ k I ) X j = A T B j , e.g. μ k = μ 0 q k .
         [ R 0 , a ] Normalize ( A T B j ( A T A + μ k I ) X i n t ) ; P 0 R 0 .
         i = 0 , σ =10tol, X j , μ k 0 = X i n t .
        while  σ > t o l and A X j , μ k i B F 2 1 4 μ k R i a F 2 < η 2 δ 2  do
            i = i + 1 .
            c = P i 1 T ( A T A + μ k I ) P i 1 1 R i 1 T R i 1 .
            X i = X i 1 + P i 1 c , X j , μ k i = X i a .
            R i = R i 1 ( A T A + μ k I ) P i + 1 c
            σ = | R i F R i 1 F | .
            d = R i 1 T R i 1 1 R i T R i .
            P i = R i + P i 1 d .
        end while
         X i n t = X μ k i .
    end while
     X ( : , j , : ) * = X j , μ k i .
end for

3.3. A preconditioned truncated tensor Conjugate-Gradient method

In this section, we consider the acceleration of Algorithm 3 by preconditioning. When the tensor M is symmetric positive definite under the t-product structure, we can get its tensor approximate Cholesky decomposition (tChol) by Algorithm 4.
Algorithm 4 Tensor Cholesky decomposition (tChol)
1:
Input:  M R m × m × n O
2:
Output:  H R m × m × n and M = H H T .
3:
M ^ fft( M ,[ ],3)
4:
for  j = 1 , 2 , . . . , n   do
5:
     H c h o l ( M ^ ( : , : , j ) ) , H is the lower triangular matrix, which is obtained by approximate Cholesky decomposition.
6:
     H ^ : , : , j H .
7:
end for
8:
H ifft( H ^ ,[ ],3).
In Algorithm 3, the coefficient tensor A T A + μ k I of the k th normal equation
( A T A + μ k I ) X = A T B
is symmetric and positive definite. We set M = A T A + μ k I and apply Algorithm 4 to obtain the decomposition of M = H H T , where each frontal slice of H is a fully sparse lower triangular matrix. After the normal equation (25) is preconditioned by M , we solve the preconditioned normal equations
A ˜ X ˜ = B ˜ ,
instead of equations (25) in Algorithm 3, where A ˜ = H 1 ( A T A + μ k I ) H T , X ˜ = H T X , B ˜ = H 1 A T B .
Applying Algorithm 3 to solve (26) instead of (25). Let X i and X ˜ i denote the solution of (25) and (26), respectively. Then we have
R ˜ i = B ˜ A ˜ X ˜ i = H 1 A T B ( H 1 ( A T A + μ k I ) H T ) H T X i
= H 1 ( A T B ( A T A + μ k I ) X i )
= H 1 R i ,
Let W i = H 1 R i , P ˜ i 1 = H T P i 1 , then we have
d ˜ = ( R ˜ i 1 T R ˜ i 1 ) 1 ( R ˜ i T R ˜ i ) = ( ( H 1 R i 1 ) T H 1 R i 1 ) 1 ( ( H 1 R i ) T H 1 R i )
= ( W i 1 T W i 1 ) 1 ( W i T W i ) ,
and
c ˜ = ( P ˜ i 1 T A ˜ P ˜ i 1 ) 1 ( R ˜ i 1 T R ˜ i 1 ) = ( ( H T P i 1 ) T H 1 ( A T A + μ k I ) H T ( H T P i 1 ) ) 1 ( ( H 1 R i 1 ) T H 1 R i 1 ) = ( ( H T P i 1 ) T H 1 ( A T A + μ k I ) P i 1 ) 1 W i 1 T W i 1 = ( P i 1 T ( A T A + μ k I ) P i 1 ) 1 W i 1 T W i 1 .
In addition, we have the iteration
X ˜ i = X ˜ i 1 + P ˜ i 1 c ˜ H T X i = H T X i 1 + H T P i 1 c ˜ X i = X i 1 + P i 1 c ˜ ,
and
R ˜ i = R ˜ i 1 A ˜ P ˜ i + 1 c ˜ H 1 R i = H 1 R i 1 H 1 A T A + μ k I H T H T P i + 1 c ˜ R i = R i 1 A T A + μ k I P i + 1 c ˜ ,
together with
P ˜ i = R ˜ i + P ˜ i 1 d ˜ H T P i = H 1 R i + H T P i 1 d ˜ P i = H T H 1 R i + P i 1 d ˜ = H T W i + P i 1 d ˜ .
Taking the preprocessing procedure (27)-(35) into Algorithm 3, we obtain the improved auto-ttCG method, which is called the truncated tensor preconditioned conjugate-gradient method of automatical determination of a suitable regularization parameter, and is abbreviated as auto-ttpCG. Algorithm 5 summarizes the auto-ttpCG method. Numerical experiments in Section show Algorighm 5 converges faster than Algorithm 3.
Algorithm 5 The auto-ttpCG method for sloving (9)
Input: A R m × m × n , B j R m × 1 × n , δ j , j = 1 , . . . , p , μ 0 , η > 1 , tol.
Output: Approximate solution X * of problem (9).
for  j = 1 , 2 , . . . p   do
     X i n t = 0 , k = 0
    while  A X j , μ k i B j F 2 > η 2 δ j 2  do
         k = k + 1 , μ k = μ 0 q k .
         H = t C h o l ( A T A + μ k I ).
         [ R 0 , a ] Normalize ( A T B j ( A T A + μ k I ) X i n t ) .
         W 0 = H 1 R 0 , P 0 = H T W 0 .
         i = 0 , σ =10tol, X j , μ k 0 = X i n t .
        while  σ > t o l and A X j , μ k i B j F 2 1 4 μ k R i a F 2 < η 2 δ 2  do
            i = i + 1 .
            c ˜ = ( P i 1 T ( A T A + μ k I ) P i 1 ) 1 W i 1 T W i 1 .
            X i = X i 1 + P i 1 c ˜ , X j , μ k i = X i a .
            R i = R i 1 A T A + μ k I P i + 1 c ˜ , W i = H 1 R i
            σ = | R i F R i 1 F | .
            d ˜ = ( W i 1 T W i 1 ) 1 ( W i T W i ) .
            P i = H T W i + P i 1 d ˜ .
        end while
         X i n t = X μ k i .
    end while
     X ( : , j , : ) * = X j , μ k i .
end for

4. Numerical Examples

This section presents three examples to show the application of Algorithms 2, 3 and 5 on the restoration of image and video. All calculations are performed in MATLAB R2018a on computers with intel core i7 and 16GB ram.
Suppose X k is the k-th approximate solution to the minimization problem (9). The quality of the approximate solution X k is defined by the relative error
Err k = X k X t r u e F X t r u e F ,
and the signal-to-noise ratio (SNR)
SNR ( X k ) = 10 log 10 X t r u e E ( X t r u e ) F 2 X k X t r u e F 2 ,
where X t r u e denotes the uncontaminated data tensor and E ( X t r u e ) is the average gray-level of X t r u e . The observed data B in (9) is contaminated by a "noise" tensor E , i.e., B = B t r u e + E . E is determined as follows. Let E j be the j th transverse slice of E , whose entries are scaled and normally distributed with a mean of zero, i.e.,
E j = ν E r , j E r , j F B t r u e , j F , j = 1 , . . . , p ,
where the data of E r , j is generated according to N(0, 1).
Example 4.1  ( Gray image ) This example considers the restoration of the blurred and noised cameraman image with the size of 256 × 1 × 256 . For the operator A , its front slices A ( : , : , i ) , i = 1 , . . . , 256 , are generated by using the MATLAB function blur, i.e.,
z = [ e x p ( ( [ 0 : b a n d 1 ] . 2 ) / ( 2 σ 2 ) ) , z e r o s ( 1 , N b a n d ) ] , A = 1 σ 2 π t o e p l i t z z 1 f l i p l r z 2 : e n d , z , A : , : , i = A i , 1 A
with N = 256 , σ = 4 and b a n d = 12 . The condition numbers of A ( i ) are c o n d ( A ( : , : , 1 ) ) = c o n d ( A ( : , : , 246 ) ) = . . . = c o n d ( A ( : , : , 256 ) ) = 11.1559 , while he condition numbers of the remaining slices are infinite. Let X t r u e denote the original undaminated cameraman image. The operator twist converts X t r u e into tensor column X t r u e R 256 × 1 × 256 for storage. The noised tensor E is generated by (36) with different noise level ν = 10 i , i = 2 , 3 . The blurred and noisy images are generated by B = A X t r u e + E .
The auto-tCG, auto-ttCG and auto-ttpCG methods are used to solve the tensor discrete linear ill-posed problems (1). The discrepancy principle is employed to determine a suitable regularization parameter by using μ k = μ 0 q k with μ 0 = A F and q = 1 2 . We set η = 1.05 in (8).
Figure 3 shows the convergence of relative errors verus (a) the iteration number k and (b) the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 corresponding in the Table 2. The iteration process is terminated when the discrepancy principle is satisfied. From Figure 3 (a), we can see that the auto-ttCG and auto-ttpCG methods do not need to solve the normal equation for all μ k ( k < 8 ) . This shows that the auto-ttCG and auto-ttpCG methods improve the auto-tCG method by the condition (24). Figure 3 (b) shows that the auto-ttpCG method converges fastest among three methods.
Table 2 lists the regularization parameter, the iteration number, the relative error, SNR and the CPU time of the optimal solution obtained by using the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise levels ν = 10 i , i = 2 , 3 . It can be seen from Table 2 that the auto-ttpCG method has the lowest relative error, highest SNR and the least CPU time for different noise level.
Figure 4 shows the reconstructed images obtained by using the auto-tCG, auto-ttCG and auto-ttpCG methods on the blurred and noised image with the noise level ν = 10 3 in Table 2. From Figure 4 we can see that the restored image by the auto-ttpCG method looks a bit better than others but the least CPU time.
Example 4.2  ( Color image ) This example shows the restoration of a blurred Lena color image by Algorithms 2, 3 and 5. The original Lena image X o r i R 256 × 256 × 3 is stored as a tensor X t r u e R 256 × 3 × 256 through the MATLAB function multi twist . We set N = 256 , σ = 3 and band=12, and get A R 256 × 256 × 256 by
z = e x p ( ( 0 : b a n d 1 . 2 ) / ( 2 σ 2 ) ) , z e r o s ( 1 , N b a n d ) ,
A = t o e p l i t z ( z ) , A : , : , i = 1 2 π σ A i , 1 A , i = 1 , . . . , 256 .
Then c o n d ( A ( : , : , 1 ) ) = . . . = c o n d ( A ( : , : , 12 ) ) = 4.68 e + 07 , and the condition number of other tensor slices of A is infinite. The noise tensor E is defined by (36). The blurred and noised tensor is derived by B = A X t r u e + E , which is shown in Figure 6 (a).
We set the color image B to be divided into multiple lateral slices and independently process each slice through (1) by using the auto-tCG, auto-ttCG and auto-ttpCG methods. Figure 5 shows the convergence of relative errors verus (a) the iteration number k and (b) the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods when dealing with the first tensor lateral slice B ( : , 1 , : ) of B with ν = 10 3 . Similar results can be derived as that in Example 5.1 from Figure 5. We can see that the auto-ttCG and auto-ttpCG methods need less iterations than the auto-tCG method from Figure 5 (a) and the auto-ttpCG method converges fastest among all methods from Figure 5 (b).
Table 3 lists the relative error, SNR and the CPU time of the optimal solution obtained by using the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise levels ν = 10 i , i = 2 , 3 . The results are very similar to that in Table 2 for different noise level.
Table 3. Example 4.2: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Table 3. Example 4.2: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Noise level Method Relative error SNR time (secs)
10 3 auto-tCG 5.90e-02 14.62 314.73
auto-ttCG 5.90e-02 14.62 262.81
auto-ttpCG 5.43e-02 15.37 103.41
10 2 auto-tCG 7.64e-02 12.37 117.48
auto-ttCG 7.48e-02 12.55 62.01
auto-ttpCG 7.01e-02 13.13 54.85
Figure 6 shows the recovered images by the auto-tCG, auto-ttCG and auto-ttpCG methods corresponding to the results with noise level ν = 10 3 . The results are very similar to that in Figure 6.
Figure 6. Example 4.2: (a) The blurred and noised Lena image and reconstructed images by (b) the auto-tCG method, (c) the auto-ttCG and (d) the auto-ttpCG method according to the noise level ν = 10 3 in Table 3.
Figure 6. Example 4.2: (a) The blurred and noised Lena image and reconstructed images by (b) the auto-tCG method, (c) the auto-ttCG and (d) the auto-ttpCG method according to the noise level ν = 10 3 in Table 3.
Preprints 88761 g006
Example 4.3 (Video) We recover the first 10 consecutive frames of blurred and noised Rhinos video from MATLAB. Each frame has 240 × 240 pixels. We store 10 pollution- and noise-free frames of the original video in the tensor X t r u e R 240 × 10 × 240 . Let z be defined by (37) with N = 240 , σ = 2 and b a n d = 12 . The coefficient tensor A is defined as follows:
A = 1 2 π σ t o e p l i t z z , A : , : , i = 1 2 π σ 2 A i , 1 A , i = 1 , . . . , 240 .
The condition number of the frontal slices of A is c o n d ( A ( : , : , i ) ) = 7.4484 e + 09 ( i 12 ) , and the condition number of the remaining frontal sections of A is infinite. The suitable regularization parameter is determined by using the discrepancy principle with η = 1.1 . The blurred- and noised tensor B is generated by B = A X t r u e + E with E R 120 × 30 × 120 being defined by (36).
Figure 7 shows the convergence of relative errors verus the iteration number k and relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods when the second frame of the video with ν = 10 3 is restored. Very similar results can be derived from Figure 7 to that in Example 5.1.
Table 4 displays the relative error, SNR and the CPU time of the optimal solution obtained by using the auto-tCG, auto-ttCG and auto-ttpCG methods for the second frame with different noise levels ν = 10 i , i = 2 , 3 . We can see that the auto-ttpCG method has the largest SNR and the lowest CPU time for different noise level ν = 10 i , i = 2 , 3 .
Figure 8 shows the original video, blurred and noised video, and the recovered video of the second frame of the video for the auto-tCG, auto-ttCG and the auto-ttpCG methods with noise level ν = 10 3 corresponding to the results in Table 4. The recovered frame by the auto-ttpCG method looks best among all recovered frames.

5. Conclusion

This paper presents three types of tensor Conjugate-Gradient methods for solving large-scale linear discrete ill-posed problems in tensor form. We first present an automatical determination strategy of a suitable regularization parameter for the tensor conjugate gradient (tCG) method. Furthermore, we develop a truncated version and a preprocessed verion of the tCG method. The proposed methods are used to different examples in image and video restoration.

Use of AI tools declaration

The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

Acknowledgements

The authors would like to thank the referees for their helpful and constructive comments. This research was supported in part by the Sichuan Science and Technology Program (grant 2022ZYD0008).

Conflict of interest

The authors declare no conflict of interest.

References

  1. Kilmer, M.E.; Martin, C.D. Factorization strategies for third order tensors. Linear Alg. Appl. 2011, 435, 641–658. [Google Scholar] [CrossRef]
  2. Hao, N.; Kilmer, M.E.; Braman, K.; Hoover, R.C. Facial recognition using tensor-tensor decompositions. SIAM J. Imaging Sci. 2013, 6, 437–463. [Google Scholar] [CrossRef]
  3. Zhang, J.; Saibaba, A. K.; Kilmer, M. E.; Aeron, S. A randomized tensor singular value decomposition based on the t-product. Numer. Linear Algebr. Appl. 2018, 25, e2179. [Google Scholar] [CrossRef]
  4. Fenu, C.; Reichel, L.; Rodriguez, G. GCV for tikhonov regularization via global Golub–Kahan decomposition. Numer. Linear Algebr. Appl. 2016, 25, 467–484. [Google Scholar] [CrossRef]
  5. Hansen, P.C. Rank-Deficient and Discrete Ill-Posed Problems; SIAM: Philadelphia, 1998. [Google Scholar]
  6. Kilmer, M. E.; Braman, K.; Hao, N.; Hoover, R. C. Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging. SIAM J. Matrix Anal. Appl. 2013, 34, 148–172. [Google Scholar] [CrossRef]
  7. Engl, H. W.; Hanke, M.; Neubauer, A. Neubauer. In Regularization of Inverse Problems; Kluwer, Dordrecht, 1996. [Google Scholar]
  8. Kindermann, S. Convergence analysis of minimization-based noise level-free parameter choice rules for linear ill-posed problems. Electron. Trans. Numer. Anal. 2011, 38, 233–257. [Google Scholar]
  9. Kindermann, S.; Raik, K. A simplified L-curve method as error estimator. Electron.Trans. Numer. Anal. 2020, 53, 217–238. [Google Scholar] [CrossRef]
  10. Reichel, L.; Rodriguez, G. Old and new parameter choice rules for discrete ill-posed problems. Numer. Algorithms 2013, 63, 65–87. [Google Scholar] [CrossRef]
  11. Reichel, L.; Ugwu, U. O. The tensor Golub–Kahan–Tikhonov method applied to the solution of ill-posed problems with at-product structure. Numer. Linear Algebr. Appl. 2022, 29, e2412. [Google Scholar] [CrossRef]
  12. Ugwu, U. O.; Reichel, L. Tensor Arnoldi–Tikhonov and GMRES-Type Methods for Ill-Posed Problems with a t-Product Structure. J. Sci. Comput. 2022, 90, 1–39. [Google Scholar]
  13. Kilmer, M. E.; Braman, K.; Hao, N.; Hoover, R. C. Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging. SIAM J. Matrix Anal. Appl. Appl. 2013, 34, 148–172. [Google Scholar] [CrossRef]
  14. Lund, K. The tensor t-function: A definition for functions of third-order tensors. Numer. Linear Algebr. Appl. 2020, 27, e2288. [Google Scholar] [CrossRef]
  15. Frommer, A.; Maass, P. Fast CG-based methods for Tikhonov–Phillips regularization. SIAM J. Sci. Comput. 1999, 20, 1831–1850. [Google Scholar] [CrossRef]
  16. Cichocki, A.; Mandic, D.; De Lathauwer, L.; Zhou, G.; Zhao, Q.; Caiafa, C.; Phan, H. A. Tensor decompositions for signal processing applications: From two-way to multiway component analysis. IEEE Signal Process. Mag. 2015, 32, 145–163. [Google Scholar] [CrossRef]
  17. Signoretto, M.; Tran Dinh, Q.; De Lathauwer, L.; Suykens, J. A. Learning with tensors: a framework based on convex optimization and spectral regularization. Mach. Learn., 2014, 94, 303-351. Mach. Learn. 2014, 94, 303–351. [Google Scholar] [CrossRef]
  18. Kilmer, M. E.; Horesh, L.; Avron, H.; Newman, E. Tensor-tensor algebra for optimal representation and compression of multiway data. Proceedings of the National Academy of Sciences. Proc. Natl. Acad. Sci. U. S. A. 2021, 118, e2015851118. [Google Scholar] [CrossRef] [PubMed]
  19. Beik, F. P. A.; Najafi–Kalyani, M.; Reichel, L. Iterative Tikhonov regularization of tensor equations based on the Arnoldi process and some of its generalizations. Appl. Numer. Math. 2020, 151, 425–447. [Google Scholar] [CrossRef]
  20. Bentbib, A. H.; Khouia, A.; Sadok, H. The LSQR method for solving tensor least-squares problems. Electron. Trans. Numer. Anal. 2022, 55, 92–111. [Google Scholar] [CrossRef]
  21. Bentbib, A. H.; El Hachimi, A.; Jbilou, K.; Ratnani, A. Fast multidimensional completion and principal component analysis methods via the cosine product. Calcolo 2022, 59, 26. [Google Scholar] [CrossRef]
  22. Khaleel, H. S.; Sagheer, S. V. M.; Baburaj, M.; George, S. N. Denoising of Rician corrupted 3D magnetic resonance images using tensor-SVD. Biomed. Signal Process. Control 2018, 44, 82–95. [Google Scholar] [CrossRef]
  23. Ugwu, U. O.; Reichel, L. Tensor regularization by truncated iteration: a comparison of some solution methods for large-scale linear discrete ill-posed problem with a t-product. arXiv 2021, arXiv:2110.02485. [Google Scholar]
  24. Zeng, C.; Ng, M. K. Decompositions of third-order tensors: HOSVD, T-SVD, and Beyond. Numer. Linear Algebr. Appl. 2020, 27, e2290. [Google Scholar] [CrossRef]
  25. El Hachimi, A.; Jbilou, K.; Ratnani, A.; Reichel, L. Spectral computation with third-order tensors using the t-product. Appl. Numer. Math. 2023, 193, 1–21. [Google Scholar] [CrossRef]
  26. Zheng, M. M.; Ni, G. Approximation strategy based on the T-product for third-order quaternion tensors with application to color video compression. Appl. Math. Lett. 2023, 140, 108587. [Google Scholar] [CrossRef]
  27. Yu, Q.; Zhang, X. (2023). T-product factorization based method for matrix and tensor completion problems. Comput. Optim. Appl. 2023, 84, 761–788. [Google Scholar] [CrossRef]
Figure 1. (a) frontal slices A ( : , : , k ) , (b) lateral slices A ( : , j , : ) and (c) tube fibers A ( i , j , : )
Figure 1. (a) frontal slices A ( : , : , k ) , (b) lateral slices A ( : , j , : ) and (c) tube fibers A ( i , j , : )
Preprints 88761 g001
Figure 2. twist-squeeze
Figure 2. twist-squeeze
Preprints 88761 g002
Figure 3. Example 4.1: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Figure 3. Example 4.1: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Preprints 88761 g003
Figure 4. Example 4.1: (a) The blurred and noised image and reconstructed images by (b) the auto-tCG method (SNR=22.36, CPU=109.87), (c) the auto-ttCG method (SNR=22.41, CPU=80.93) and (d) the auto-ttpCG method (SNR=22.48, CPU=33.98) according to the noise level ν = 10 3 in Table 2.
Figure 4. Example 4.1: (a) The blurred and noised image and reconstructed images by (b) the auto-tCG method (SNR=22.36, CPU=109.87), (c) the auto-ttCG method (SNR=22.41, CPU=80.93) and (d) the auto-ttpCG method (SNR=22.48, CPU=33.98) according to the noise level ν = 10 3 in Table 2.
Preprints 88761 g004
Figure 5. Example 4.2: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Figure 5. Example 4.2: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Preprints 88761 g005
Figure 7. Example 4.3: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Figure 7. Example 4.3: Comparison of convergence between (a) relative errors verus the iteration number k and (b) relative errors verus the CPU time for the auto-tCG, auto-ttCG and auto-ttpCG methods with the noise level ν = 10 3 .
Preprints 88761 g007
Figure 8. Example 4.3: (a) Original image, (b) the blurred and noisy image and recovered images by (c) the auto-tCG method, (d) the auto-ttCG and (e) the auto-ttpCG method according to the noise level ν = 10 3 in Table 4.
Figure 8. Example 4.3: (a) Original image, (b) the blurred and noisy image and recovered images by (c) the auto-tCG method, (d) the auto-ttCG and (e) the auto-ttpCG method according to the noise level ν = 10 3 in Table 4.
Preprints 88761 g008
Table 1. Description of notations
Table 1. Description of notations
Notation Interpretation
A T transpose of tensors
A 1 inverse of tensor, A T = ( A 1 ) T = ( A T ) 1
A ^ FFT of A along the third mode
unfold ( A ) the block column matrix of A
bcirc ( A ) the block-circulant matrix
I identity tensor
A matrix
I identity matrix
A F the Frobenius norm of tensors A , i.e, A F = i = 1 l j = 1 m k = 1 n a i j k 2 .
t-product
A j , A ( : , j , : ) the jth tensor column of A , jth lateral slice of A
A ( : , : , j ) the jth frontal slice of tensor A
d tube
A , B A , B = i j k a i j k b i j k
A , B A , B = i k a i 1 k b i 1 k
Table 2. Example 4.1: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Table 2. Example 4.1: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Noise level Method k μ k Relative error SNR CPU (secs)
10 3 auto-tCG 15 1.96e-05 3.54e-02 22.36 109.87
auto-ttCG 15 1.96e-05 3.52e-02 22.41 80.93
auto-ttpCG 15 1.96e-05 3.49e-02 22.48 33.98
10 2 auto-tCG 11 3.14e-04 8.74e-02 14.51 81.94
auto-ttCG 11 3.14e-04 8.64e-02 14.61 26.42
auto-ttpCG 11 3.14e-04 8.54e-02 14.72 18.50
Table 4. Example 4.3: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Table 4. Example 4.3: Comparison of relative error, SNR, and CPU time between the auto-tCG, auto-ttCG and auto-ttpCG methods with different noise level ν = 10 i , i = 2 , 3 .
Noise level Method Relative error SNR time (secs)
10 3 auto-tCG 2.94e-02 23.17 697.78
auto-ttCG 2.92e-02 23.23 487.35
auto-ttpCG 2.66e-02 24.05 214.16
10 2 auto-tCG 5.24e-02 18.15 480.75
auto-ttCG 5.10e-02 18.38 281.54
auto-ttpCG 4.74e-02 19.02 156.44
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated