Xue 2019
Xue 2019
https://doi.org/10.1007/s11760-019-01415-6
ORIGINAL PAPER
Abstract
Recently, total variation regularization has become a standard technique, and even a basic tool for image denoising and
deconvolution. Generally, the recovery quality strongly depends on the regularization parameter. In this work, we develop a
recursive evaluation of Stein’s unbiased risk estimate (SURE) for the parameter selection, based on specific reconstruction
algorithms. It enables us to monitor the evolution of mean squared error (MSE) during the iterations. In particular, to deal
with large-scale data, we propose a Monte Carlo simulation for the practical computation of SURE, which is free of any
explicit matrix operation. Experimental results show that the proposed recursive SURE could lead to highly accurate estimate
of regularization parameter and nearly optimal restoration performance in terms of MSE.
Keywords Total variation · Denoising · Deconvolution · Stein’s unbiased risk estimate (SURE) · Jacobian recursion
1 Introduction
N
2 2
TV(x) = (D1 x)n + (D2 x)n +α (3)
n=1
Problem statement—Consider the standard image recovery
problem: find a good estimate of original image x0 ∈ R N where D1 and D2 denote the horizontal and vertical first-order
from the following degradation model [15,20,22]: differences, respectively. The parameter α = 0 corresponds
to the standard TV definition. We use the smooth approx-
y = Hx0 + (1)
imation with small α > 0, since it simplifies numerical
computations due to the differentiability [9].
where y ∈ R N is the observed image, H ∈ R N ×N denotes
TV is particularly effective for recovering those signals
the observation matrix, which represents either identity for
with piecewise constant region while preserving edges [12].
denoising or convolution for deconvolution, and ∈ R N is an
Recently, people extended the basic TV-norm to more gen-
additive white Gaussian noise with known variance σ 2 > 0.
eral form of ϕ(Dx2 ) that models a priori of the first-order
Since the seminal work of ROF [12], total variation (TV)
gradients of an image [13]. Here, ϕ is a potential, possibly
regularization has become a standard technique [17,19]:
non-convex, function. It is reduced to the standard TV-norm
1 when ϕ(t) = t. In this work, we focus on the TV mini-
xλ = arg min
Hx − y2 + λ · TV(x) (2)
2 mization and attempt to find a proper value of λ for a good
2
x
L(x)
restoration quality. This work may help to gain some insights
into more complicated function of ϕ.
Here, λ is a regularization parameter, which is essential for Related works—There have been a number of criteria for this
the recovery quality of
xλ . The isotropic TV term is defined selection of λ, for example:
as [9]:
– Generalized cross-validation [5]: It is often used for linear
This work was supported by the National Natural Science Foundation
of China (No. 61401013). estimates, not applicable for the nonlinear reconstruction
considered here.
B Feng Xue – L-curve method [7]: This procedure is not fully auto-
[email protected] mated and often requires hand tuning or selection.
1 National Key Laboratory of Science and Technology on Test
Physics and Numerical Mathematics, Beijing 100076, China
123
Signal, Image and Video Processing
– Discrepancy principle [8]: This criterion is easy to com- 2 Recursive evaluation of SURE for TV
pute and however may cause a loss of restoration quality.1 denoising
In this paper, we quantify the restoration performance by Now, we consider image denoising problem, i.e. H = I in
the mean squared error (MSE) [1,22]: (2). To perform the SURE-based selection of λ, we need to
compute the solution
xλ and its SURE.
1 2
MSE = E xλ − x0 2 (4)
N 2.1 Basic scheme of Chambolle’s algorithm [2]
and attempt to select a value of λ, such that the corresponding
solution xλ achieves minimum MSE. Many algorithms can be used to find the TV solution xλ , e.g.
Notice that the MSE is inaccessible due to the unknown [3,4,6,11,14]. Here, we apply a dual-based iterative projec-
x0 . In practice, Stein’s unbiased risk estimate (SURE) has tion algorithm—Chambolle’s algorithm—to solve (2), since
been proposed as a statistical substitute for MSE [1,16]: it is one of the most popular TV minimization solvers and
has been extensively used in the recent decade. The original
1 2 form of the Chambolle’s iteration was described in [2]. Now,
SURE =
xλ 2 − 2yT H−T
xλ + 2σ 2 Tr H−T Jy (
xλ )
N we rewrite the algorithm in matrix language:
1 2
+ x0 2 (5) ⎛
w(i)
⎞
N
(i) ⎜ (i) τ ⎟
u(i+1) =V ⎜ ⎝u − λ D (y + λDT u(i) )⎟ (6)
since it depends on the observed data y only.2 Tr in (5) denotes ⎠
xλ ) ∈ R N ×N is a Jacobian matrix
the matrix trace. Here, Jy ( x(i)
defined as [21,23]:
where τ is a step size and the gradient operator D is D =
xλ )m
∂(
Jy (
xλ ) m,n = [DT1 , DT2 ]T ∈ R2N ×N [D1 and D2 are the same as in (3)].
∂ yn The vector u(i) ∈ R2N lives in D-domain of an image, which
can be transformed back to the image domain by DT u. The
The statistical unbiasedness of SURE w.r.t. true MSE has (i)
been proved in [22]. Recently, SURE has become a popular diagonal matrix V is:
criterion for parameter selection, in the context of nonlinear (i)
(i) V 0
denoising/deconvolution [1,22], and 1 -based sparse recov- V = ∈ R2N ×2N
ery [18,23]. However, to our best knowledge, there are very 0 V(i)
few researches on the application of SURE to TV-based
reconstruction, which is the purpose of this paper. where the diagonal block V(i) ∈ R N ×N is given by:
123
Signal, Image and Video Processing
First, we split (6) into two parts—vertical and horizontal Noting that x(i) = y + λDT u(i) in (6), we have:
differences:
(i) (i)
(i+1)
(i) Jy (x(i) ) = I + λDT1 Jy u1 + λDT2 Jy u2 (11)
u1 V(i) 0 w1
(i+1) = 0 V (i) (i)
u2 w2
Thus, the Jacobian matrix Jy (x(i) ) can be evaluated in this
(i)
u(i+1) V w(i) recursive manner, until the convergence of Chambolle’s iter-
ation, summarized in Algorithm 1.
(i+1)
From (6), the Jacobian matrix of u1 is:
(10) (13)
123
Signal, Image and Video Processing
Following the similar procedure with Sect. 2, we now con- 3.2 Jacobian recursion of ADMM
sider the TV deconvolution problem, where the SURE also
requires to compute the solution
xλ and its SURE. For the deconvolution problem, the SURE needs to compute
H−1 in (5). However, it is observed that for the ill-conditioned
3.1 Basic scheme of ADMM matrix H, the simple inversion H−1 may cause numerical
To find
xλ , we choose a typical alternating direction method 3We can see that the Chambolle’s iteration is readily incorporated into
of multipliers (ADMM) for TV deconvolution [i.e. H being ADMM, see Sect. 2.1 for details.
123
Signal, Image and Video Processing
instability of SURE [22]. Hence, we use the regularized input given y and H
inverse Hβ−1 to replace H−1 : initial z(0)
Chambolle
Hβ−1 = (HT H + βI)−1 HT update x(i) compute z(i)
by (15) by (16)
compute SURE
with a parameter β. The regularized SURE becomes: by (19), (17)
(i) (i)
update nx compute nz
1 by (20) by (20) ADMM
SURE = x(i) 2 − 2yT H−T x(i) + 2σ 2 Tr H−T Jy (x(i) )
N 2 β β ADMM iteration by i := i + 1
1 2
+ x0 2 (17)
N output xλ and SURE
Refer to [18,22] for the similar treatment. Fig. 1 SURE-MC evaluation for ADMM (Chambolle’s algorithm is
By the similar derivations with Sect. 2.2, we obtain the for obtaining z(i) )
Jacobian recursions for ADMM as:
⎧
⎪ Jy (x(i) ) = (HT H + μI)−1 HT + μJy (z(i) ) 4 Experimental results and discussion
⎪
⎪
⎨ J (u(i, j+1) ) = V(i, j) J (u(i, j) ) − τ P(i, j) J (z(i, j) )
y 1 y 1 λ 1 y
(i, j+1) (i, j) (i, j) (18) 4.1 Experimental setting
⎪
⎪ Jy (u2 ) = V(i, j) Jy (u2 ) − τλ P2 Jy (z(i, j) )
⎪
⎩ J (z(i, j) ) = J (x(i) ) + λ DT J (u(i, j) )
y y μ y
The test dataset contains four 8-bit images of size 256 × 256
(i, j) (i, j) (i, j) (i, j) (i, j)
or 512 × 512 displayed in Fig. 2, covering a wide range of
where P1 = (W1 C1 + V(i, j) )D1 + W1 C2 D2 natural images.
(i, j) (i, j) (i, j) (i, j) (i, j)
and P2 = W2 C1 D1 + (W2 C2 + V(i, j) )D2 . For both denoising and deconvolution, we always termi-
(i, j) (i, j) (i, j) (i, j) nate the iterative algorithms, when the relative error of the
Here, W1 , W2 , C1 and C2 are defined similarly
with Sect. 2.2. objective value L(x(i) ) in (2) is below 10−5 .
The restoration performance is measured by the peak
3.3 Monte Carlo for SURE evaluation signal-to-noise ratio (PSNR), defined as (in dB) [20,22]:
Similar to Sect. 2.3, we adopt Monte Carlo to evaluate the 2552
PSNR = 10 × log10
trace term of SURE as:
x − x0 22 /N
⎧ ⎫
⎪
⎪ ⎪
⎪
−T ⎨ ⎬ We choose α = 10−12 in the TV definition of (3) and set
(i) T −T (i)
Tr Hβ Jy (x ) = E n0 Hβ Jy (x )n0 (19) the parameter τ = 1/4 in (6), as suggested in [2].
⎪
⎪ ⎪ ⎪
⎩ (i)
⎭
nx
4.2 Image denoising
with input white Gaussian noise n0 ∼ N (0, I N ). Then,
multiplying n0 on both sides of (18), we obtain the noise 4.2.1 SURE evaluation for Chambolle’s algorithm
evolution during ADMM:
⎧ (i) (i)
First, we need to verify the accuracy of SURE w.r.t. MSE for
⎪
⎪ nx = B−1 HT n0 + μB−1 nz the Chambolle’s iteration. Figure 3 shows the convergence
⎪
⎨ n(i, j+1) (i, j) (i, j) (i, j)
u1 = V(i, j) nu1 − τλ P1 nz of Chambolle and evolution of SURE, under the noise levels
(i, j+1) (i, j) (i, j) (i, j) (20)
⎪
⎪ n = V(i, j) nu2 − τλ P2 nz of σ 2 = 1, 10 and 100, respectively. We can see that: (1)
⎪ u2
⎩ (i, j) (i, j) (i, (i, j)
nx + μλ DT1 nu1 + μλ DT2 nu2
j)
nz =
123
Signal, Image and Video Processing
objective value
objective value
7.00E+007
0.01 750000
95500 0.01 0.01
1E-3
6.00E+007
1E-3 1E-3
1E-4 700000
95000 1E-4 1E-4
1E-5 5.00E+007
evolutions of SURE and MSE of x(i) during Chambolle’s iteration denoised images by SURE
1.00 MSE
10
MSE 100
PSNR=32.72dB PSNR=36.19dB PSNR=29.20dB PSNR=24.76dB
SURE SURE
0.98 9 90
MSE/SURE
MSE/SURE
MSE/SURE
0.96
8 80
0.94
0.92 7 70
0.90
60
MSE
6
SURE
0.88
1 2 3 4 5 4 8 12 16 20 15 30 45 60 75
iteration number iteration number iteration number
MSE/SURE
MSE/SURE
90
1.6 16
1.4
opt. λ = 0.21 by SURE min.
opt. λ = 0.21 by MSE min. 12
opt. λ = 1.39 by SURE min.
opt. λ = 1.39 by MSE min.
80
opt. λ = 4.29 by SURE min.
opt. λ = 4.29 by MSE min. ues of λ and obtain Fig. 4, where the optimal λ is easy to
70
1.2
1.0 8
60 MSE
recognize.
SURE
0.8
1E-3 0.01 0.1 1 0.01 0.1 1 10 0.1 1 10
regularization parameter regularization parameter regularization parameter
objective value
objective value
1.5E6
.
9E5 1E-4
1E-4 1E-4
3E6 6E7
Nσ2
1E-5 1E-5 1E-5
6E5
50 100 150 50 100 150 200 250
(i+1)
λ(i)
50 100 150 200
iteration number iteration number
λ =
iteration number
update of λ
update of λ
6
2.0
18
which finally satisfies the discrepancy condition when con-
4
λ(i) → 8.37 1.6
λ(i) → 2.42 12 λ(i) → 29.55
2
1.2 6
verged. Figure 5 shows a few examples of the parameter
50 100
iteration number
150 200
0.8
50 100
iteration number
150 50 100 150
iteration number
200 250 update by DP.
Table 1 shows the complete comparisons between the pro-
Fig. 5 The convergence of Chambolle with the parameter update by posed SURE-based method and discrepancy principle. Here,
discrepancy principle
‘DP’ denotes discrepancy principle. The format of this table
est. λ , where the upper value is the selected value of λ by
is PSNR
the objective value keeps decreasing until convergence; (2) DP/SURE/MSE and the lower one is its resultant denoising
SURE is always close to MSE during the iterations. PSNR (in dB) using the corresponding λ. The symbol ‘–’
123
Signal, Image and Video Processing
(1) Cameraman
Rational, BSNR=40dB
(2) Coco
Uniform, BSNR=30dB
(3) Bridge
Gaussian, BSNR=10dB indicates that the method fails to find an optimal λ for this
objective value of L(x(i) ) — converegence of ADMM
120000
objective value
10
450000 objective value
1 9.00E+007
objective value 0.1
case. The MSE is not accessible in practice and thus shown
110000
error of obj. value 1
420000
error of obj. value
0.1
error of obj. value
in italics. It is the comparison benchmark, indicating the best
objective value
objective value
0.1 7.50E+007
100000 390000 0.01
90000
0.01
360000 1E-3
1E-3 PSNR performance we can achieve. We can see that com-
1E-3 6.00E+007
80000
1E-4
330000
1E-4
1E-4
pared to the MSE minimization, the PSNR by DP is worse
300000
70000
5 10 15 20 25 30 35 40
1E-5
10 20 30 40 50
1E-5 4.50E+007
50 100 150
iteration number
200 250
1E-5
than optimal PSNR by 1 dB in average, whereas the SURE
iteration number iteration number
160
evolutions of SURE and MSE of x(i) during ADMM iteration minimization yields negligible PSNR loss (within 0.02 dB).
600
MSE MSE MSE
140
SURE
100 SURE
550
SURE
Figure 6 shows a number of visual examples.
MSE/SURE
MSE/SURE
MSE/SURE
120 80
500
100
80
60
450
400
4.3 Image deconvolution
40
60 350
5 10 15 20 25 30 35 40 10 20 30 40 50 50 100 150 200 250
iteration number iteration number iteration number
4.3.1 Experimental setting
Fig. 7 The convergence of ADMM and evolution of SURE for fixed λ:
(1) λ = 0.1, (2) λ = 1, (3) λ = 10 For deconvolution problem, we consider the following
benchmark convolution kernels commonly used in [10,22]:
(1) Cameraman (2) Coco (3) Bridge
Rational, BSNR=40dB Uniform, BSNR=30dB Gaussian, BSNR=10dB
i, j = −7, . . . , 7;
MSE/SURE
MSE/SURE
MSE/SURE
600
300
20
1E-3 0.01 0.1 1 0.01 0.1 1 0.1 1 10
regularization parameter regularization parameter regularization parameter
– 9 × 9 uniform blur;
2+ j 2
Fig. 8 The global optimization of λ for TV deconvolution by ADMM – Gaussian kernel h(i, j) = C · exp − i 2s 2 with
s = 2.0.
(1) Cameraman (2) Coco (3) Bridge
Separable, BSNR=30dB Uniform, BSNR=20dB Gaussian, BSNR=10dB
/
where C is a normalization factor, s.t. i, j h(i, j) = 1.0.
1 0.14
4.5E5 3.28E7
9.0E5
objective value
objective value
objective value
update of λ
0.12
3.24E7
λ(i) → 0.51
3.0E5 0.1 7.5E5
6.0E5
0.1
3.20E7 0.1
Gaussian noise with various variance σ 2 , corresponding to
3.16E7
1.5E5
10 20 30
objective value
update of λ
40 50
0.01
20 40 60
objective value
update of λ
80 100
0.01
5 10 15 20
objective value
update of λ
25 30
0.08 blur signal-to-noise ratio (BSNR) being 40, 30, 20 and 10 dB,
iteration number iteration number iteration number
123
Signal, Image and Video Processing
observed images
PSNR=22.25dB PSNR=26.33dB PSNR=29.59dB PSNR=20.49dB 4.3.2 SURE evaluation for ADMM
Nσ2
λ(i+1) = λ(i)
y − Hx(i) 22
123
Signal, Image and Video Processing
5 Conclusions 11. Ramirez, C., Argaez, M.: An 1 minimization algorithm for non-
smooth regularization in image processing. Signal Image Video
Process. 9, 373–386 (2015)
In this paper, we presented a SURE-based method for 12. Rudin, L.I., Osher, S., Fatemi, E.: Nonlinear total variation based
automatically tuning regularization parameter for TV-based noise removal algorithms. Physica D 60, 259–268 (1992)
recovery. In particular, we proposed a recursive evaluation 13. Selesnick, I., Parekh, A., Bayram, I.: Convex 1-D total variation
and Monte Carlo simulation for the practical computation. denoising with non-convex regularization. IEEE Signal Process.
Lett. 22(2), 141–144 (2015)
Numerical results showed the superior performance of SURE 14. Shen, C., Bao, X., Tan, J., Liu, S., Liu, Z.: Two noise-robust axial
to other criteria for parameter selection, e.g. discrepancy scanning multi-image phase retrieval algorithms based on Pauta
principle. criterion and smoothness constraint. Opt. Express 25(14), 16235–
This proposed method, in principle, can be extended 16249 (2017)
15. Siadat, M., Aghazadeh, N., Öktem, O.: Reordering for improving
to more complicated (possibly non-convex) regularizers
global Arnoldi–Tikhonov method in image restoration problems.
[13,18,23]. Future work will also deal with the SURE- Signal Image Video Process. 12, 497–504 (2018)
based multiple parameter selection and faster optimization 16. Stein, C.M.: Estimation of the mean of a multivariate normal dis-
of SURE, to accelerate the global search used here. tribution. In: Stein, C.M. (ed.) The Annals of Statistics, vol. 9, No.
6, pp. 1135–1151 (1981)
17. Tao, M., Yang, J.: Alternating Direction Algorithms for Total Varia-
tion Deconvolution in Image Reconstruction. Optimization Online,
TR0918, Department of Mathmatics, Nanjing University (2009)
18. Vonesch, C., Ramani, S., Unser, M.: Recursive risk estimation for
References non-linear image deconvolution with a wavelet-domain sparsity
constraint. In: IEEE International conference on Image processing,
1. Blu, T., Luisier, F.: The SURE-LET approach to image denoising. pp. 665–668 (2008)
IEEE Trans. Image Process. 16(11), 2778–2786 (2007) 19. Wang, Y., Yang, J., Yin, W., Zhang, Y.: A new alternating mini-
2. Chambolle, A.: An algorithm for total variation minimization and mization algorithm for total variation image reconstruction. SIAM
applications. J. Math. Imaging Vis. 20, 89–97 (2004) J. Imaging Sci. 1(3), 248–272 (2008)
3. Chen, D.: Inext alternating direction method based on Newton 20. Xue, F., Blu, T.: A novel SURE-based criterion for parametric PSF
descent algorithm with application to Poisson image deblurring. estimation. IEEE Trans. Image Process. 24(2), 595–607 (2015)
Signal Image Video Process. 11, 89–96 (2017) 21. Xue, F., Du, R., Liu, J.: A recursive predictive risk estimate for
4. Dell’Acqua, P.: ν acceleration of statistical iterative methods for proximal algorithms. In: Proceedings of the 41st IEEE International
image restoration. Signal Image Video Process. 10, 927–934 (2016) Conference on Acoustics, Speech and Signal Processing, Shanghai,
5. Golub, G., Heath, M., Wahba, G.: Generalized cross-validation as a China, March 20–25, pp. 4498–4502 (2016)
method for choosing a good ridge parameter. Technometrics 21(2), 22. Xue, F., Luisier, F., Blu, T.: Multi-Wiener SURE-LET deconvolu-
215–223 (1979) tion. IEEE Trans. Image Process. 22(5), 1954–1968 (2013)
6. Guo, C., Li, Q., Wei, C., Tan, J., Liu, S., Liu, Z.: Axial multi-image 23. Xue, F., Yagola, A.G., Liu, J., Meng, G.: Recursive SURE for iter-
phase retrieval under tilt illumination. Sci. Rep. 7, 7562 (2017) ative reweighted least square algorithms. Inverse Probl. Sci. Eng.
7. Hansen, P.C.: Analysis of discrete ill-posed problems by means of 24(4), 625–646 (2016)
the L-curve. SIAM Rev. 34(4), 561–580 (1992)
8. Morozov, V.: Methods for Solving Incorrectly Posed Problems.
Springer, New York (1984)
Publisher’s Note Springer Nature remains neutral with regard to juris-
9. Osher, S., Burger, M., Goldfarb, D., Xu, J., Yin, W.: An iterative
dictional claims in published maps and institutional affiliations.
regularization method for total variation-based image restoration.
SIAM J. Multiscale Model. Simul. 4(2), 460–489 (2005)
10. Pan, H., Blu, T.: An iterative linear expansion of thresholds for 1 -
based image restoration. IEEE Trans. Image Process. 22(9), 3715–
3728 (2013)
123