Tomography
Tomography
tomography reconstruction
Zenith Purisha1 , Carl Jidling2 , Niklas Wahlström2 , Thomas B.
Schön2 , Simo Särkkä1
arXiv:1809.03779v3 [cs.CV] 3 Jul 2019
1
Department of Electrical Engineering and Automation, Aalto University, Finland
2
Department of Information Technology, Uppsala University, Sweden
E-mail: [email protected]
2019
1. Introduction
instance, the medical imaging community. The limited data case—also referred to as
sparse projections—calls for a good solution for several important reasons, including:
• the needs to examine a patient by using low radiation doses to reduce the risk of
malignancy or to in vivo samples to avoid the modification of the properties of
living tissues,
• geometric restrictions in the measurement setting make it difficult to acquire the
complete data [12], such as in mammography [13, 14, 15, 16] and electron imaging
[17],
• the high demand to obtain the data using short acquisition times and to avoid
massive memory storage, and
• the needs to avoid—or at least minimize the impact of—the moving artifacts during
the acquisition.
Classical algorithms—such as FBP—fail to generate good image reconstruction
when dense sampling is not possible and we only have access to limited data. The
under-sampling of the projection data makes the image reconstruction (in classical
terms) an ill-posed problem [18]. In other words, the inverse problem is sensitive to
measurement noise and modeling errors. Hence, alternative and more powerful methods
are required. Statistical estimation methods play an important role in handling the ill-
posedness of the problem by restating the inverse problem as a well-posed extension in
a larger space of probability distributions [19]. Over the years there have been a lot of
work on tomographic reconstruction from limited data using statistical methods (see,
e.g., [14, 20, 21, 22, 23, 24]).
In the statistical approach, incorporation of a priori knowledge is a crucial part
in improving the quality of the image reconstructed from limited projection data.
That can be viewed as an equivalent of the regularization parameter in classical
regularization methods. However, statistical methods, unlike classical regularization
methods, also provide a principled means to estimate the parameters of the prior
(i.e., the hyperparameters) which corresponds to automatic tuning of regularization
parameters.
In our work we build the statistical model by using a Gaussian process model [25]
with a hierarchical prior in which the (hyper)parameters in the prior become part of
the inference problem. As this kind of hierarchical prior can be seen as an instance
of a Gaussian process (GP) regression model, the computational methods developed
for GP regression in machine learning context [25] become applicable. It is worth
noting that some works on employing GP methods for tomographic problems have
also appeared before. An iterative algorithm to compute a maximum likelihood point
in which the prior information is represented by GP is introduced in [26]. In [27, 28],
tomographic reconstruction using GPs to model the strain field from neutron Bragg-edge
measurements has been studied. Tomographic inversion using GP for plasma fusion and
soft x-ray tomography have been done in [29, 30]. Nevertheless, the proposed approach
is different from the existing work.
Probabilistic approach to limited-data computed tomography reconstruction 3
dI(s)
= −f (x1 (s), x2 (s))ds, (1)
I(s)
and by integrating both sides of (1), the following relationship is obtained
Z R
I0
f (x1 (s), x2 (s))ds = log , (2)
−R Id
where dxL denotes the 1-dimensional Lebesgue measure along the line defined by
L = {(x1 , x2 ) ∈ R2 : x1 cos θ + x2 sin θ = r}, where θ ∈ [0, π) is the angle and r ∈ R is
the distance of L from the origin as shown in Figure 1.
The parametrization of the straight line L with respect to the arc length s can be
written as:
x1 (s, θ, r) = r cos(θ) − s sin(θ),
(4)
x2 (s, θ, r) = r sin(θ) + s cos(θ).
In this work, the object is placed inside a circular disk with radius R. Then, as a
function of r and θ the line integral in (3) can be written as
Z R
Rf (r, θ) = f (x1 (s, θ, r), x2 (s, θ, r)) ds
−R
Z R (5)
0
= f (x + sû)ds,
−R
where
h iT h iT
0
x = r cos(θ) r sin(θ) , û = − sin(θ) cos(θ) .
approximation for the attenuation data in (2) can be used [32, 33]. Recall that a
logarithm of the intensity is involved in (5), and so additive noise is a reasonable model
for the electronic noise.
We collect a set of measurements as
Z R
yi = f (x0i + sûi )ds + εi , (6)
−R
where i corresponds to the data point index. The corresponding inverse problem is given
the noisy measurement data {yi }ni=1 in (6) to reconstruct the object f .
The GP is uniquely specified by the mean function m(x) = E[f (x)] and the covariance
function k(x, x0 ) = E[(f (x) − m(x))(f (x0 ) − m(x0 ))]. The mean function encodes our
prior belief of the value of f in any point. In lack of better knowledge it is common to
pick m(x) = 0, a choice that we will stick to also in this paper.
The covariance function on the other hand describes the covariance between two
different function values f (x) and f (x0 ). The choice of covariance function is the most
important part in the GP model, as it stipulates the properties assigned to f . A few
different options are discussed in Section 2.4.
As data is collected our belief about f is updated. The aim of regression is to
predict the function value f (x∗ ) at an unseen test point x∗ by conditioning on the seen
data. Consider direct function measurements on the form
yi = f (xi ) + εi , (8)
where εi is independent and identically distributed (iid) Gaussian noise with variance
σ 2 , that is, εi ∼ N (0, σ 2 ). Let the measurements be stored in the vector y. Then the
mean value and the variance of the predictive distribution p(f (x∗ ) | y) are given by [25]
Here the vector k∗ contains the covariances between f (x∗ ) and each measurement while
the matrix K contains the covariance between all measurements, such that
Figure 2. Left: GP prediction (shaded surface) obtained from the measurements (red
stars, also indicated by their deviation from the prediction). Right: slice plot of the
blue line in the left figure, including the 95% credibility region.
In general we can not expect closed form solutions to (14a)–(14b) and numerical
computations are then required. However, even with efficient numerical methods,
the process of selecting the hyperparameters is tedious since the hyperparameters are
in general not decoupled from the integrand and the integrals need to be computed
repeatedly in several iterations. In this paper, we avoid this by using the basis function
expansion that will be described in Section 2.6.
where again r = x − x0 .
The perhaps most commonly used covariance function within the machine learning
context [25] is the squared exponential (SE) covariance function
2 1 2
kSE (r) = σf exp − 2 krk2 , (16)
2l
which has the following spectral density
l2 kωk22
SSE (ω) = σf2 (2π)d/2 ld exp − , (17)
2
where d is the dimensionality of x (in our case d = 2). The SE covariance
function is characterized by the magnitude parameter σf and the length scale l. The
squared exponential covariance function is popular due to its simplicity and ease of
implementation. It corresponds to a process whose sample paths are infinitely many
times differentiable and thus the functions modeled by it are very smooth.
Another common family of covariance functions is given by the Matérn class
1−ν
√ !ν √ !
2 2νkrk 2 2νkrk2
kMatern (r) = σf2 Kν , (18a)
Γ(ν) l l
d d/2
−(ν+d/2)
Γ(ν + d/2)(2ν)ν 2ν
22 π 2
SMatern (ω) = σf + kωk2 , (18b)
Γ(ν)l2ν l2
where Kν is a modified Bessel function [25]. The smoothness of the process is increased
with the parameter ν: in the limit ν → ∞ we recover the squared exponential covariance
function.
Probabilistic approach to limited-data computed tomography reconstruction 8
Gaussian processes are also closely connected to classical spline smoothing [35] as
well as other classical regularization methods [19, 36] for inverse problems. Although
the construction of the corresponding covariance function is hard (or impossible), it
is still possible to construct the corresponding spectral density in many cases. With
these spectral densities and the basis function method of Section 2.6, we can construct
probabilistic versions of the classical regularization methods as discussed in the next
section.
σf2
S(ω) = . (20)
F[L∗ L]
where S denotes the spectral density of the covariance function, and m is the truncation
number. The basis functions φi (x) and eigenvalues λi are obtained from the solution to
the Laplace eigenvalue problem on the domain Ω
(
−∆φi (x) = λi φi (x), x ∈ Ω,
(24)
φi (x) = 0, x ∈ ∂Ω.
The entries Φij can be computed in closed form with details given in Appendix A. Now
we substitute Q ≈ ΦT ΛΦ and q∗ ≈ ΦT Λφ∗ to obtain
When using the spectral densities corresponding to the classical regularization methods
in (21) and (22), the mean equation reduces to the classical solution (on the given
basis). However, also for the classical regularization methods we can compute the
variance function which gives uncertainty estimate for the solution which in the classical
formulation is not available. Furthermore, the hyperparameter estimation methods
outlined in the next section provide principled means to estimate the parameters also
in the classical regularization methods.
Probabilistic approach to limited-data computed tomography reconstruction 10
3. Hyperparameter estimation
In this section, we will consider some methods for estimating the hyperparameters. The
free parameters of the covariance function, for example, the parameters σf and l in
the squared exponential covariance function, are together with the noise parameter σ
referred to as the hyperparameters of the model. In this work, we employ a Bayesian
approach to estimate the hyperparameters, and comparisons with standard parameter
estimation methods such as L-curve and cross-validation methods are given as well.
where Q(σf , l) is defined by (14a). The posterior distribution of parameters can now be
written as follows:
3.4. Cross-validation
As a comparison, we also consider to use methods of cross-validation (CV) for model
selection. In k-fold CV, the data are partitioned into k disjoint sets yj , and at each
round j of CV, the predictive likelihood of the set yj is computed given the rest of
the data y−j . These likelihoods are used to monitor the predictive performance of the
model. This performance is used to estimate the generalization error, and it can be used
to carry out model selection [40, 25, 41].
The Bayesian CV estimate of the predictive fit with given parameters ϕ is
n
X
CV = log p(yj | y−j , ϕ), (31)
j=1
where p(yj | y−j , ϕ) is the predictive likelihood of the data yj given the rest of the data.
The best parameter values with respect to CV can be computed by enumerating the
possible parameter values and selecting the one which gives the best fit in terms of CV.
4. Experimental results
In this section, we present numerical results using the GP model for limited x-
ray tomography problems. All the computations were implemented in Matlab 9.4
(R2018a) and performed on an Intel Core i5 at 2.3 GHz and CPU 8GB 2133MHz
LPDDR3 memory.
For both simulated data (see Section 4.1) and real data (see Section 4.2) we use
m = 104 basis functions in (23). The measurements are obtained from the line integral of
each x-ray over the attenuation coefficient of the measured objects. The measurements
are taken for each direction (angle of view), and later they will be referred to as
projections. The same number of rays in each direction is used. The computation
of the hyperparameters is carried out using the Metropolis–Hastings algorithms with
5 000 samples, and the first 1 000 samples are thrown away (burn-in period). The
reconstruction is computed by taking the conditional mean of the object estimate.
attenuation function values. The measurements (i.e. sinogram) of the chest phantom
are computed using the radon command in Matlab and corrupted by additive white
Gaussian noise with zero mean and 0.1 variance (σtrue = 0.32).
Several reconstructions of the chest phantom using different covariance functions,
namely squared exponential (SE), Matérn, Laplacian, and Tikhonov, are presented. For
the SE, Matérn, and Laplacian covariance functions, the paramameters σf , l, and σ are
estimated using the proposed method. We use ν = 1 for the Matérn covariance. As
for the Tikhonov covariance, it is not characterized by the length scale l, and hence
only σf and σ are estimated. All the estimated parameters are reported in Table 2.
Figure 3 presents the histograms of the 1-d marginal posterior distribution of each
parameters using different covariance functions. The histograms show the distribution
of the parameters samples in the Metropolis–Hastings samples. The results show that
the σf estimate for SE and Matérn covariances is 0.12, while for Laplacian and Tikhonov,
the estimates are 0.05 and 0.64. For Matérn, Laplacian, and Tikhonov covariance
functions, the σ estimates are concentrated around the same values 0.34 − 0.39 with
standard deviation (SD) between 0.02 − 0.03. These noise estimates are well-estimated
the ground-truth noise, σtrue = 0.32, with the absolute error is between 0.02 − 0.07. The
estimate of the SE kernel appears to overestimate the noise, σ = 0.60. It is reported that
the length-scale parameter, l, for Laplacian and SE covariance functions are concentrated
in the same values, while for Matérn yields higher estimate, l = 10.14.
Figure 4(c)-(f) shows GP reconstructions of the 2D chest phantom using different
covariance functions from 9 projections (uniformly spaced) out of 180◦ angle of view
and 185 number of rays for each projection. The computation times for all numerical
tests are reported in Table 1. The Metropolis–Hastings reconstruction shows longer
computational time due to the need for generation of a large number of samples from
the posterior distribution. However, the benefit of this algorithm is that it is easy to
implement and it is reliable for sampling from high dimensional distributions.
The numerical test of the simulated data reconstructions is compared against figures
of merit, namely:
• the relative error (RE)
kftrue − frec k2
,
kftrue k2
peakval2
10 log10 ,
MSE
where peakval is the maximum possible value of the image and MSE is the mean
square error between ftrue and frec ,
as shown in Table 3.
In practice, image quality in CT depends on other parameters as well, such as image
contrast, spatial resolution, and image noise [43]. These parameters can be evaluated
when the CT device is equipped with CT numbers for various materials, high-resolution
image is available, and statistical fluctuations of image noise which require several times
of measurement to record random variations in detected x-ray intensity are acquired.
However, in this work, the collected datasets are not supported by the aforementioned
factors and they fall outside the scope of this paper. The results presented here are
focusing on the implementation of a new algorithm to limited-data CT reconstruction
and are reported as a preliminary study.
Reconstruction using a conventional method is computed as well with the built-
in Matlab function iradon, which uses the FBP to invert the Radon transform. It
reconstructs a two-dimensional slice of the sample from the corresponding projections.
The angles for which the projections are available are given as an argument to the
function. Linear interpolation is applied during the backprojection and a Ram–Lak
or ramp filter is used. The FBP reconstruction of the chest phantom is shown in
Figure 4(b). For comparison, FBP reconstructions computed using some other filters
are seen in Figure 5.
Table 2. The GP parameter estimates for the chest phantom. The estimates are
calculated using the conditional mean, and the standard deviation (SD) values are
also reported in parentheses.
We also compared the results to the L-curve method and the CV:
• The L-curve method is applied to the Laplacian and the Tikhonov covariances
and the L-curve plots from different values of parameter 10−1 ≤ σ ≤ 10 for both
covariances are shown in Figure 6. Both plots show that the corner of the L-curve
is located in between 0.2 ≤ σ ≤ 1.
Probabilistic approach to limited-data computed tomography reconstruction 14
σf l σ
160
100 100
140
80 120 80
SE
100
60 60
80
40 60 40
40
20 20
20
0 0 0
0 0.2 0.4 0.6 0.8 1 0 10 20 30 40 50 0 0.2 0.4 0.6 0.8 1
70 60 70
60 60
Matérn
50
50 50
40
40 40
30
30 30
20
20 20
10
10 10
0 0 0
0 0.2 0.4 0.6 0.8 1 0 10 20 30 40 50 0 0.2 0.4 0.6 0.8 1
40 70 70
35 60
Laplacian
60
30
50 50
25
40 40
20
30 30
15
20 20
10
5 10 10
0 0 0
0 0.2 0.4 0.6 0.8 1 0 10 20 30 40 50 0 0.2 0.4 0.6 0.8 1
60 50
45
50
40
Tikhonov
35
40
30
30 25
20
20
15
10
10
5
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
(a) (b)
(c) (d)
Figure 5. Filtered backprojection reconstructions using (a) Shepp–Logan filter, (b)
Cosine filter (c) Hamming filter, (d) Hann filter. Values of relative error (RE) are
between 23.6 − 25.2 and PSNR values are between 18.1 − 19.9%.
Probabilistic approach to limited-data computed tomography reconstruction 16
40
σ = 0.1 180 σ = 0.1
35 160
140
30
120
kfσ (x)k2
kfσ (x)k2
25
100
20 0.2 ≤ σ ≤ 1
80
σ = 10
1
15
≤
60
σ
10
≤
40
σ = 102
2
σ = 10
0.
5 20
σ = 102
0 0
0 100 200 300 400 500 600 5 10 15 20 25 30 35 40
(a) (b)
Figure 6. The L-curve for (a) Tikhonov and (b) Laplacian covariance from the chest
phantom reconstruction.
• The CV is tested for the Laplacian and Tikhonov covariances using point-wise
evaluation of 10−2 ≤ σ ≤ 1 and 10−2 ≤ σf ≤ 1. For the Laplacian covariance,
several points of length scale 1 ≤ ` ≤ 100 are tested as well. The minimum
prediction error was obtained for σf = 0.8, σ = 0.8 and ` = 10. For the Tikhonov
covariance, the minimum prediction error was obtained for σ = 0.5 and σf = 0.5.
The estimates of σf and σ for Laplacian are 0.8 and 0.5, respectively, and they
give the same estimates for the Tikhonov covariance function. The estimates of σ
for both kernels appear to overestimate the σtrue . The absolute error is in between
0.18 − 0.48. The length-scale estimate from Laplacian covariance, l = 10, appears
Probabilistic approach to limited-data computed tomography reconstruction 17
(d) (e)
Figure 7. (a) A ground truth of 2D chest phantom. (b) & (c) reconstructions using
L-curve parameter choice method with Laplacian (using σ = 1) and Tikhonov (using
σ = 0.2) covariance functions, respectively. (d) & (e) reconstructions using CV with
Laplacian and Tikhonov covariance functions, respectively
4.3. Discussion
We have presented x-ray tomography reconstructions from both simulated and real
data for limited projections (i.e. sparse sampling) using an approach based on
Probabilistic approach to limited-data computed tomography reconstruction 18
Table 4. Estimated GP parameters for the carved cheese using Matérn covariance
function. The estimates are calculated using the conditional mean, and the standard
deviation (SD) values are also reported in parentheses.
Figure 8. (a) FBP reconstruction (Ram–Lak filter) of the carved cheese using dense
360 projections. (b) Filtered backprojection reconstruction from 15 projections. (c)
GP reconstruction using Matérn covariance from 15 projections.
the Gaussian process. However, other limited-data problems such as limited angle
tomography could be explored as well. The quality of GP reconstructions using different
covariance functions looks rather the same qualitatively. However, quantitatively, the
reconstruction using Matérn covariance is the best one: it has the lowest RE 23.26%
and the highest PSNR 22.76. PSNR describes the similarity of the original target
with the reconstructioned image (the higher value, the better of the reconstruction).
Figures of merit estimates are not available for the real cheese data since there is
no comparable ground truth. Nevertheless, the quality of the reconstruction can be
observed qualitatively by comparing with the FBP reconstruction obtained with dense
360 projections from 360 degrees shown in Figure 8(a). The corresponding parameter
estimates for the chest phantom and the cheese are reported in Table 2 and 4. For the
Probabilistic approach to limited-data computed tomography reconstruction 19
chest phantom case, the estimate of parameter σ using Matérn, Laplacian and Tikhonov
kernels tend to be close to the true value σtrue . As for the SE covariance, the standard
deviation of noise is overestimated.
The reconstructions produced by the FBP benchmark algorithm using sparse
projections are overwhelmed by streak artefacts due to the nature of backprojection
reconstruction, as shown in Figure 4(b) for the chest phantom and Figure 8(b) the for
cheese target. The edges of the target are badly reconstructed. Due to the artefacts,
especially for the chest phantom, it is difficult to distinguish the lighter region (which is
assumed to be tissue) and the black region (air). The FBP reconstruction has the worst
quality and it is confirmed in Table 3 that it has a high RE value (25.86%) and the
lowest PSNR (18.44). FBP reconstructions computed with different filters are shown in
Figure 5. However, there is no significant improvement in the images as it is clarified
by the RE and PSNR values in the caption as well as by qualitative investigation.
On the other hand, the GP reconstructions outperform the FBP algorithm in terms
of image quality as reported in the figures of merit. The PSNR values of the GP-
based reconstructions are higher than that of the FBP reconstruction. Nevertheless,
in GP reconstructions, sharp boundaries are difficult to achieve due to the smoothness
assumptions embedded in the model.
The GP prior clearly suppresses the artifacts in the reconstructions as shown in
Figure 4(c) and 8(c). In Figure 4(c), the air and tissue region are recovered much better,
since the prominent artefacts are much less. In Figure 8(c), the air region (outside the
cheese and the C and T letters) are much sharper than in the FBP reconstruction.
Overall, the results indicate that the image quality can be improved significantly by
employing the GP method.
In Figure 7 the image reconstructions using L-curve and CV methods are presented.
The quality of the reconstructions is reported in Table 3 as well. In these methods, finer
point-wise evaluations might help to improve the quality of the reconstructions.
We emphasize that in the proposed GP-approach, some parameters in the prior
is a part of the inference problem (see Equation (16)). Henceforth, we can avoid the
difficulty in choosing the prior parameters. This problem corresponds to the classical
regularization methods, in which selecting the regularization parameters is a very crucial
step to produce a good reconstruction.
5. Conclusions
Acknowledgments
Authors acknowledge support from the Academy of Finland (314474 and 313708).
Furthermore, this research was financially supported by the Swedish Foundation for
Strategic Research (SSF) via the project ASSEMBLE (contract number: RIT15-0012).
Probabilistic approach to limited-data computed tomography reconstruction 21
Here we derive the closed-form expression of the entries Φij stated in (26b). We get that
Z R
Φij = φi (x0j + sûj )ds
−R
Z R
1
= √ sin(ϕi1 rj cos θj − ϕi1 s sin θj + ϕi1 L1 ) sin(ϕi2 rj sin θj + ϕi2 s cos θj + ϕi2 L2 )ds
L1 L2 −R
Z R
1
= √ sin(αij s + βij ) sin(γij s + δij )ds
L1 L2 −R
Z R
1
= √ cos((αij − γij )s + βij − δij ) − cos((αij + γij )s + βij + δij )ds
2 L1 L1 −R
1 h 1 1 iR
= √ sin((αij − γij )s + βij − δij ) − sin((αij + γij )s + βij + δij )
2 L1 L1 αij − γij αij + γij −R
1 1 1
= √ sin((αij − γij )R + βij − δij ) − sin((αij + γij )R + βij + δij )
2 L1 L1 αij − γij αij + γij
1 1
− sin(−(αij − γij )R + βij − δij ) + sin(−(α + γij )R + βij + δij ) ,
αij − γij αij + γij
(A.1)
where
References
[1] A. M. Cormack, “Representation of a function by its line integrals, with some radiological
applications,” Journal of Applied physics, vol. 34, no. 9, pp. 2722–2727, 1963.
[2] A. M. Cormack, “Representation of a function by its line integrals, with some radiological
applications. ii,” Journal of Applied Physics, vol. 35, no. 10, pp. 2908–2913, 1964.
[3] G. T. Herman, “Image reconstruction from projections,” Topics in Applied Physics, vol. 32, 1979.
[4] P. Kuchment, The Radon transform and medical imaging. SIAM, 2013.
[5] N. R. Council et al., Mathematics and physics of emerging biomedical imaging. National Academies
Press, 1996.
[6] L. A. Shepp and J. Kruskal, “Computerized tomography: the new medical x-ray technology,”
American Mathematical Monthly, pp. 420–439, 1978.
[7] S. Akin and A. Kovscek, “Computed tomography in petroleum engineering research,” Geological
Society, London, Special Publications, vol. 215, no. 1, pp. 23–38, 2003.
[8] L. Cartz, Nondestructive testing. ASM International, 1995.
[9] L. De Chiffre, S. Carmignato, J.-P. Kruth, R. Schmitt, and A. Weckenmann, “Industrial
applications of computed tomography,” CIRP Annals-Manufacturing Technology, vol. 63, no. 2,
pp. 655–677, 2014.
Probabilistic approach to limited-data computed tomography reconstruction 22
[10] A. C. Kak and M. Slaney, Principles of computerized tomographic imaging. IEEE press New York,
1988.
[11] T. M. Buzug, Computed tomography: from photon statistics to modern cone-beam CT. Springer
Science & Business Media, 2008.
[12] N. Riis, J. Frøsig, Y. Dong, and P. Hansen, “Limited-data x-ray CT for underwater pipeline
inspection,” Inverse Problems, vol. 34, no. 3, p. 034002, 2018.
[13] L. T. Niklason, B. T. Christian, L. E. Niklason, D. B. Kopans, D. E. Castleberry, B. Opsahl-Ong,
C. E. Landberg, P. J. Slanetz, A. A. Giardino, R. Moore, et al., “Digital tomosynthesis in breast
imaging,” Radiology, vol. 205, no. 2, pp. 399–406, 1997.
[14] M. Rantala, S. Vanska, S. Jarvenpaa, M. Kalke, M. Lassas, J. Moberg, and S. Siltanen, “Wavelet-
based reconstruction for limited-angle x-ray tomography,” IEEE transactions on medical
imaging, vol. 25, no. 2, pp. 210–217, 2006.
[15] T. Wu, A. Stewart, M. Stanton, T. McCauley, W. Phillips, D. B. Kopans, R. H. Moore, J. W.
Eberhard, B. Opsahl-Ong, L. Niklason, et al., “Tomographic mammography using a limited
number of low-dose cone-beam projection images,” Medical physics, vol. 30, no. 3, pp. 365–380,
2003.
[16] Y. Zhang, H.-P. Chan, B. Sahiner, J. Wei, M. M. Goodsitt, L. M. Hadjiiski, J. Ge, and C. Zhou, “A
comparative study of limited-angle cone-beam reconstruction methods for breast tomosynthesis,”
Medical physics, vol. 33, no. 10, pp. 3781–3795, 2006.
[17] D. Fanelli and O. Öktem, “Electron tomography: a short overview with an emphasis on the
absorption potential model for the forward problem,” Inverse Problems, vol. 24, no. 1, p. 013001,
2008.
[18] F. Natterer, The mathematics of computerized tomography, vol. 32. Siam, 1986.
[19] J. Kaipio and E. Somersalo, Statistical and computational inverse problems, vol. 160. Springer
Science & Business Media, 2006.
[20] C. A. Bouman and K. Sauer, “A unified approach to statistical tomography using coordinate
descent optimization,” IEEE Transactions on image processing, vol. 5, no. 3, pp. 480–492, 1996.
[21] H. Haario, A. Kallonen, M. Laine, E. Niemi, Z. Purisha, and S. Siltanen, “Shape recovery for sparse-
data tomography,” Mathematical Methods in the Applied Sciences, vol. 40, no. 18, pp. 6649–
6669, 2017.
[22] V. Kolehmainen, S. Siltanen, S. Järvenpää, J. P. Kaipio, P. Koistinen, M. Lassas, J. Pirttilä,
and E. Somersalo, “Statistical inversion for medical x-ray tomography with few radiographs: II.
application to dental radiology,” Physics in Medicine & Biology, vol. 48, no. 10, p. 1465, 2003.
[23] S. Siltanen, V. Kolehmainen, S. Järvenpää, J. Kaipio, P. Koistinen, M. Lassas, J. Pirttilä, and
E. Somersalo, “Statistical inversion for medical x-ray tomography with few radiographs: I.
general theory,” Physics in Medicine & Biology, vol. 48, no. 10, p. 1437, 2003.
[24] K. Sauer, J. Sachs, and C. Klifa, “Bayesian estimation of 3-D objects from few radiographs,” IEEE
transactions on nuclear science, vol. 41, no. 5, pp. 1780–1790, 1994.
[25] C. E. Rasmussen and C. K. I. Williams, Gaussian processes for machine learning. MIT press,
Cambridge, MA, 2006.
[26] A. Tarantola, Inverse problem theory and methods for model parameter estimation, vol. 89. siam,
2005.
[27] J. Hendriks, A. Gregg, C. Wensrich, and W. Wills, “Implementation of traction constraints in
bragg-edge neutron transmission strain tomography,” arXiv preprint arXiv:1805.09760, 2018.
[28] C. Jidling, J. Hendriks, N. Wahlström, A. Gregg, T. B. Schön, C. Wensrich, and A. Wills,
“Probabilistic modelling and reconstruction of strain,” arXiv preprint arXiv:1802.03636, 2018.
[29] D. Li, J. Svensson, H. Thomsen, F. Medina, A. Werner, and R. Wolf, “Bayesian soft x-ray
tomography using non-stationary Gaussian processes,” Review of Scientific Instruments, vol. 84,
no. 8, p. 083506, 2013.
[30] J. Svensson, Non-parametric tomography using Gaussian processes. EFDA, 2011.
[31] A. Solin and S. Särkkä, “Hilbert space methods for reduced-rank Gaussian process regression,”
Probabilistic approach to limited-data computed tomography reconstruction 23