# Efficient implementation of a local tomography reconstruction algorithm

- Pierre Paleo
^{1, 2}Email authorView ORCID ID profile and - Alessandro Mirone
^{2}

**3**:5

https://doi.org/10.1186/s40679-017-0038-1

© The Author(s) 2017

**Received: **2 September 2016

**Accepted: **9 January 2017

**Published: **19 January 2017

## Abstract

We propose an efficient implementation of an interior tomography reconstruction method based on a known subregion. This method iteratively refines a reconstruction, aiming at reducing the local tomography artifacts. To cope with the ever increasing data volumes, this method is highly optimized on two aspects: firstly, the problem is reformulated to reduce the number of variables, and secondly, the operators involved in the optimization algorithms are efficiently implemented. Results show that \(4096^2\) slices can be processed in tens of seconds, while being beyond the reach of equivalent exact local tomography method.

### Keywords

Tomography Interior problem Local tomography Reconstruction algorithm## Background

Computed tomography is a permanently evolving X-ray imaging technique finding various applications from medical imaging to materials science and non-destructive testing [1]. From a series of radiographs acquired at various angles, the interior of the scanned volume is reconstructed. In the ideal case, i.e., with a sufficient signal-to-noise ratio and a proper modeling, the reconstruction can be computed relatively easily. However, experimental constraints usually move away from the ideal case and require more advanced reconstruction methods. Among these constraints is the imaging of an object bigger than the detector field of view. This setup is called *local tomography* or region-of-interest (ROI) tomography.

*interior problem*. The interior problem has infinitely many solutions in general, in the sense that a solution can differ from another solution by an infinitely differentiable function [3].

Local tomography methods basically consist in estimating the exterior of the ROI from the acquired measurements. This can be done with sinogram extrapolation (see for example [4, 5]) or in the slice domain. These methods, although they can yield satisfactory results, are only heuristics in general. Solutions computed with these methods often suffer from the *cupping effect*, which is an artifact appearing as a low-frequency bias.

Theoretical investigations, however, found that, with a prior knowledge on region of interest, the interior problem can be solved [6]. This prior knowledge can be about the values of a subregion of the ROI or about the nature of the solution [7, 8].

In this work, we consider a reconstruction method described in [9] where a prior knowledge is available as values of a subregion. We show how the reduction of the number of unknowns can be coupled with an efficient implementation of the involved operators, in order to cope with the scales of modern datasets.

## Methods

Based on the observation that the filtered backprojection with extrapolation provides satisfactory reconstruction of medium and high frequencies of the slice, the method aims at improving the reconstructed slice by removing the local tomography artifacts visible as low-frequency artifacts (cupping effect). This correction is performed by representing the reconstruction error in a coarse basis, reducing the number of degrees of freedom of the problem.

From an exact iterative reconstruction method, the reconstruction problem is reformulated to incorporate the local tomography setup, the prior knowledge constraint and the representation of the image in a coarse basis. Each operator of the forward model is analyzed to enable an efficient implementation. Notably, the projector is reduced to a point-projector which is efficiently implemented with a sparse matrix-vector multiplication.

The local reconstruction implementation is validated on simulated data for which the cupping effect is prominent. The proposed method is compared against another exact local reconstruction method also based on known region. Two criteria are compared: the number of required iterations to achieve an acceptable reconstruction, and the total execution time. The former reflects the relative ill-posedness of the problem and the performance of the chosen optimization algorithm, while the latter shows how the efficient implementation of the operators affects the reconstruction time. The benchmarks are carried on data compatible with modern data volumes, up to \(4096^2\) pixels with 4000 projections.

## An iterative correction algorithm for local tomography

### Local tomography and artifacts

The most common local tomography reconstruction method is extrapolating the sinogram before computing the filtered backprojection (FBP), hereafter denoted *padded FBP*. The extrapolation is usually done by replicating the sinogram boundary values. This prevents truncation artifact (Gibbs phenomenon) from occurring, and often provides acceptable results [10].

However, this technique can fail when the ROI is surrounded by anisotropic and/or strongly absorbing material or when the reconstruction has intrinsically low contrast (for example different parts with the same linear absorption coefficient).

The notable local tomography artifact is the cupping effect. On a reconstructed image, local tomography artifacts appear as a varying contrast. The gray values are typically higher far from the center than close to the center, forming a “cup.” The cupping is also visible when plotting an image line passing through the center, as a function of the pixel location. Such lines are hereby called *profiles*, for example, the vertical line profile is the vertical line of the image passing through the center.

In this work, we examine a family of exact reconstruction methods based on a known subregion. We implement a method handling a reduced number of unknowns by expressing the image in a coarse basis in order to correct the cupping effect.

### Iterative reconstruction

*P*is the model of the projection operator,

*d*the acquired data, and

*x*is the unknown volume to recover. In the remainder of this paper, we consider reconstruction of a single slice rather than a volume, so

*x*shall denote two-dimensional slices. In parallel beam geometry, as it is the case in synchrotrons, reconstruction can be performed by reconstructing the slices independently.

In this context, the reconstructed slice *x* is an image of support \(N \times N = N^2\), where *N* is the number of pixels of the detector horizontally. The sinogram *d* support is \(N \times N_p\), where \(N_p\) is the number of projections. Thus, the projector is theoretically an operator of dimensions \((N \times N_p, \, N^2)\), assuming that slices are stacked as one-dimensional \(N^2\) vectors, and sinograms are stacked as one-dimensional \(N \times N_p\) vectors.

*d*, as explained in the introductory part. In order for the solution to be acceptable, the exterior of the ROI has to be estimated. This can be done by extending the support of

*x*to iteratively estimate the exterior by solving (3)

*d*), the size of the projected solution has to be consistent with the acquired data. Thus, the projection is cropped by the means of an operator

*C*to recover the original local geometry. The cropping operator

*C*maps an extended sinogram of support \(N_2 \times N_p\) to a sinogram of support \(N \times N_p\), by keeping only the

*N*central columns. This models the truncation in the local tomography setup, where the detector is not large enough to image the entire object support \(N_2\). In practice, the cropping operation is implemented inside the projector \(\tilde{P}\) by simply restricting the projection to the detector limited field of view

*N*. In the formulas, the cropping operator

*C*is explicitly separated from the projector \(\tilde{P}\) to highlight the local setup in the forward model.

Efficient implementations of the projection and backprojection operators enable to solve problem (3). The ASTRA toolbox [11], for example, has versatile geometry capabilities and built-in algorithms for solving (3) for \(\phi (x) = 0\).

In this work, we consider the case where a subregion is known. This prior knowledge on the volume can be used to constrain the sets of solutions. A uniqueness theorem was stated in [6] along with a reconstruction algorithm based on differentiated backprojection and projection onto convex sets to invert the finite Hilbert transform. This algorithm, however, is difficult to implement, and no implementation is readily available for experiments.

We focus on a simpler approach based on formalism (3). In this formulation, the prior knowledge can be encoded in several ways. The first is to enforce the values of \(\tilde{x}\) in the known region, for example, using an indicator function. The second is to add a term penalizing the distance between the values of \(\tilde{x}\) in the known region and the actual values. We adopt the latter approach, which was proposed, for example, in [12].

Let \(\Omega\) denote the domain where the values of the volume are known. It is a subset (possibly a union of subsets) of the image support \(N^2\), and we denote \(N_\Omega\) its cardinality, that is, the total number of known pixels. Let \(x_{|\Omega }\) denote the values of *x* inside the known region. The prior knowledge is encoded by \(\phi (x) = \lambda \left\| x_{|\Omega } - u_0 \right\| _2^2\), where \(u_0\) denotes the known values inside \(\Omega\) and \(\lambda \ge 0\) is a parameter weighting the fidelity to the known zone. Both \(x_{|\Omega }\) and \(u_0\) have \(N_\Omega\) components.

This approach, however, has two drawbacks. The first is using a prior which might not be accurate: in this example, Total Variation promotes piecewise-constant images and is thus not adapted for complex samples. The second drawback is on the computational side. Adding a non-differentiable prior involves to change the optimization algorithm for another probably less efficient in the sense that more iterations are required to reach convergence. In the examples, the preconditioned Chambolle–Pock algorithm described in [14] was used for the TV minimization. Approximatively, 3000 iterations are required to approximately get rid of the cupping effect (when approximatively 500 are required in the case of a complete scan), and more than 10,000 iterations are required to get the line profiles shown in Fig. 8. This approach is impracticable for modern datasets with increasing amount of data: on the one hand, projection and backprojection become costly operations, while on the other hand, even more iterations are required due to the higher number of variables.

The main contribution of this work is an efficient implementation of the method described in [9]. The method is based on the following observation: the padded FBP reconstruction yields acceptable reconstruction of features of the ROI [15], but can suffer from a low-frequency bias (cupping effect). On the other hand, iterative algorithms converge slowly due to the high indeterminacy of the problem, even with a known subregion. For these reasons, a refinement of the initial reconstruction is computed rather than the complete solution.

### Correction of the low-frequency bias

#### Estimating the reconstruction error

Let \(x_0\) be a reconstruction of the region of interest with the padded FBP technique and \(x^\sharp\) be the true values of the region of interest. Both are slices of support \(N^2\) pixels. The reconstruction error, unknown in practice, is denoted \(e = x^\sharp - x_0\). This error mainly consists in low-frequency artifacts (the cupping effect).

*C*is a truncation operator. As the initial reconstruction is constant, problem (4) can be rewritten as

*f*. As the reconstruction error in the ROI is \(e = x^\sharp - x_0\), we can write

*e*) in the ROI, not in the whole volume (\(\tilde{e}\)). Since the projection of

*e*is different from the cropped projection of \(\tilde{e}\), the term \(d - P x_0\) only approximates the projection of the reconstruction error in the ROI. This quantity is nevertheless used as an approximation of the projection of the reconstruction error in the ROI. Once the optimal correction term \(\hat{x_e}\) is found, the resulting reconstruction is simply computed as \(x = \tilde{C}(\tilde{x_0} + \hat{x_e})\) where \(\tilde{C}\) is a cropping operator in the image domain, mapping images of support \(N_2^2\) to images of support \(N^2\).

#### Reducing the degrees of freedom

The principle of the implemented method is to refine an initial solution of the local tomography problem, knowing that middle and high-frequency features are usually well recovered. By focusing on the low frequencies, the complexity of problem (3) can be reduced by solving a simpler problem. Complexity reduction is achieved by expressing the reconstruction error in a coarse basis.

*e*is estimated by \(\hat{e}\) as a convolution between a finite discrete Dirac comb and a two-dimensional Gaussian function \(g_\sigma\) defined by Eq. (7)

*s*is the spacing (in pixels) between points of the Dirac comb. The summation in (8) actually occurs on a finite support. In our implementation, the Gaussian function is truncated at \(3\sigma\) at each side, so the sum takes place on a \(\lfloor 6\sigma +1\rfloor \times \lfloor 6\sigma +1\rfloor\) pixels square.

*G*denote the operator mapping the coefficients \(c_{i,j}\) to the image \(\hat{e}\) through convolution formula (8). The coefficients vector

*c*is estimated by solving Problem (9)

Thus, Problem (9) is solved instead of Problem (3). In Problem (9), the unknowns are the coefficients *c* of the coarse basis. As there are much less coefficients *c* in the coarse representation than pixels in the extended image support \(N_2^2\), the degrees of freedom is accordingly reduced.

Solving (9) requires the computation of the operators *C*, \(\tilde{P}\), *G*, and possibly their adjoints. The implementation of the crop operator *C* is straightforward, as it consists in truncating the sinogram to the size of the acquired data. In practice, it consists in modifying the projector \(\tilde{P}\) so that the projections are limited to the reduced detector field of view \(N_2\). The operator *G* can be described as follows. Coefficients \(c_{u,v}\) are placed every \(s > 0\) pixel on an image of the size of the extended reconstruction \(\tilde{x_0}\). This image (a two-dimensional Dirac comb in the continuum case) is then convolved by the kernel \(g_\sigma\). Lastly, an efficient implementation of the projection and backprojection operators is needed to solve (9). This is discussed in the implementation section.

#### Adding the known zone constraint

We now describe how the known zone constraint is implemented in formalism (9). In work [9], the knowledge available as known zone values in the slice is translated in the coarse representation basis: a subset of Gaussian coefficients is fitted to values in the known zone \(\Omega\); these coefficients are then used as a constraint for the reconstruction.

## High-performance implementation

After having reduced the number of degrees of freedom for problem (4), we describe an efficient implementation of the involved operators based on look-up tables (LUTs).

### Projection a of Gaussian tiling

The choice of a Gaussian basis for a coarse representation of the correction term is based on a characteristic of the Gaussian kernel: it is both rotationally invariant and separable [16]. These two properties provide a computational advantage: the order of projection and convolution can somehow be exchanged.

More precisely, given an image *y* consisting of the Gaussian coefficients evenly placed with a spacing *s*, the standard way to compute \(\tilde{P} G y\) is first performing the convolution *Gy* defined by (8) and then projecting with \(\tilde{P}\). An equivalent computation, however, can be done by first projecting the image of isolated points *y*, and then convolving each line of the resulting sinogram by a one-dimensional Gaussian function. This is illustrated in Fig. 9.

*point-projector*based on look-up tables.

### LUT-based point-projector

As previously discussed, the operators involved in forward model (10) are a cropping operator, a one-dimensional convolution, and a projector. The convolution can be efficiently implemented, either in the Fourier space or in the direct space when one of the functions has a small support. Therefore, a fast projector is essential for solving (10) in an iterative fashion. In our case, the object to project has a very special structure, as it consists in points spaced by several pixels. Thus, standard projectors of tomography softwares can be replaced by a more efficient implementation, hereby called *point- projector*, based on look-up tables.

In the remainder, the following notations are used: The support of the original image is \(N^2\). The number of projections is \(N_p\), so the acquired sinogram has size \(N \times N_p\). The size of the extended image is \(N_2 ^2\) where \(N_2 \ge N\). The number of Gaussian functions used to tile the support is \(N_g\). The spacing between Gaussian blobs on the image is *s* ; thus we have \(N_g \simeq \left( \frac{N_2}{s} \right) ^2\) in a first approximation. We also use the following indexes convention : Gaussian coefficients are numbered with \(i \in [0, N_g]\), and sinogram indexes are numbered with \(k \in [0, N_s]\) where \(N_s = N_2 \times N_p\) is the size of the (extended) sinogram.

*J*is built so that for each

*i*,

*J*[

*i*] is the “list” of locations in sinogram hit by this point after projection. The LUT

*J*is an array of size \(N_g \times N_p\). Each entry \(J_{i,j}\) corresponds to a position, in the sinogram, that is hit by a projected point \(i \in [0, N_g]\). For example, entry \(J_{0,2}\) is an index in the sinogram that is hit by point 0 ; and entry \(J_{5,j}\) are an indexes hit by point 5 for all

*j*. This is illustrated in Figs. 10, 11.

When computing the sinogram, however, the look-up table *J* is best accessed “backward”: for a given position \(k \in [0, N_s[\) in the sinogram, we have to determine which points are hitting it through projection. To this end, two look-up tables *J* and \(\mathrm {Pos}\) are built. For \(k \in [0, N_s]\), \(\mathrm {Pos}[k]\) indicates a position in LUT *J*, and \(J[p_k]\) is a coefficient number \(i \in [0, N_g]\) being projected at position *k*. Therefore, the LUT *J* does not contain sinogram indexes anymore, but rather coefficient indexes. This is illustrated in Fig. 12. The LUT *J* is re-ordered such that the interval \([p_k, p_{k+1}-1]\) gives access to an indexes range in *J* ; this index range is the set of all coefficients indexes being projected on sinogram index *k*.

The point-projector is described by Algorithm 1. The matrix *W*, indexed in the same way as *J*, contains the weights of the projections: depending on the position of a point in the image and the projection angle, its projection does not exactly fall into a sinogram pixel. The matrix *W* thus encodes the geometric contribution of the projection of the points.

*Compressed Sparse Row*(CSR) format [17], where LUT

*J*corresponds to “col_ind,” LUT \(\mathrm {Pos}\) corresponds to “row_ptr,” and matrix

*W*contains the values. Storing the entire “linear-algebra” projection matrix without compression would entail to store \((N_2^2)\times (N_2 \cdot N_p)\) elements, which is impracticable (for example, more than one terabyte is required for a \(1024^2\) slice). However, as each slice point is projected on at most \(N_p\) sinogram positions, this matrix actually has at most \(N_2^2 \times N_p\) non-zero elements. Additionally, as the slice is reduced on a coarse basis, there are \(\left( \frac{N_2}{s} \right) ^2 \times N_p\) non-zero values to store in this case. The format described above is used to store these elements. Algorithm 1 is thus no more than a matrix-vector multiplication with a sparse matrix in CSR format.

This approach for computing the point-projector is friendly in a memory-write point of view: after accumulating the contributions of all coefficients projected on position *k*, the sinogram at index *k*, \(\mathrm {sino}[k]\), is updated accordingly. This is especially important for GPU implementation, as consecutive threads access contiguous memory locations, which is a coalesced access pattern. On GPUs, each memory transaction actually entails accessing *L* bytes, so coalesced access to 32 bits scalars results in a read or write of *L* / 4 addresses in a single transaction (for example, \(L = 128\) for modern NVidia GPUs).

### Implementation of the adjoint operators

As a gradient-based optimization algorithm is used for solving (10), the adjoint of operator \(C \tilde{P} G\) has to be computed. This operator \(G^T \tilde{P}^T C^T\) consists in extending the sinogram with zeros, point-backprojecting and retrieving the Gaussian components from the backprojected image. As mentioned above, the operator *G* can be described as \(G = H_\sigma U\) where *U* is an upsampling operator (here with a factor *s*), and \(H_\sigma\) is the convolution with 2D Gaussian kernel (7). Thus, \(G^T = H_\sigma ^T U^T\) which is a downsampling followed by a convolution with kernel (7). The actual computation is then \(G^T \tilde{P}^T C^T = H_\sigma ^T U^T \tilde{P}^T C^T = U^T \tilde{P}^T H_\sigma ^1 C^T\) where \(H_\sigma ^1\) is a one-dimensional convolution on the sinogram rows.

As previously, these operations can be merged. As \(G^T \tilde{P}^T C^T\) returns a Gaussian coefficients vector from a sinogram, only the coefficients are of interest here. Therefore, the point-backprojector \(\tilde{P}^T\) is merged with the downsampling \(U^T\) as previously. For a given coefficient, we have to find which sinogram entries backproject on the coefficient position. This approach avoids to compute useless \(N_g \times (s-1)^2\) backprojections points on the image, as it is downsampled afterward.

*W*.

### Parallel implementation

In modern experiments carried on X-ray light sources, the data volumes, produced by new generations of detectors, always overwhelm the computing power. Simply waiting for more powerful machines is of little hope, as advances in detectors overrun the Moore’s law. Instead, an algorithmic work has to be accomplished to exploit parallelism of modern architectures. In the last decade, the advent of general-purpose GPU (GPGPU) computing was advantageously used, especially in tomography.

The proposed method has been implemented in the PyHST2 software [18] used at ESRF for tomographic reconstruction, with the CUDA language targeting Nvidia GPUs. The point-projector and point-backprojector, which are the most time-consuming operators, are implemented as efficient CUDA kernels. As for Algorithms 1 and 2, the CUDA point-projector and point-backprojector are implemented as matrix-vector multiplication with a matrix in CSR format.

We describe here the implementation of the point-projector, i.e., the computation of the sinogram values \(\mathrm {sino}[k]\) for \(k \in [0, N_s]\). The point-backprojector follows the same principle. To compute the sinogram value \(\mathrm {sino}[k]\), the LUT *J* has to be accessed from \(p_k\) to \(p_{k+1}-1\) as illustrated in Fig. 13. This memory range is accessed in parallel by threads of the many-cores GPU with the following principle. Each thread reads \(m \ge 1\) values in the LUT. With these values *J*[*j*], where \(j = p_k, p_k +1, \ldots\), the coefficients vector is accessed at \(\mathrm {coeffs}[J[j]]\). The threads are grouped in blocks, and each thread updates a temporary array in shared memory with the contributions read in \(\mathrm {coeffs}[J[j]]\). Then, in each block, the shared array is accumulated by one thread. The result is added to \(\mathrm {sino}[k]\). This is illustrated in Fig. 14.

*J*, as it is the biggest data structure of the method. As it has been re-arranged so that the interval \([\mathrm {pos}[k], \mathrm {pos}[k+1]-1]\) is a contiguous memory range in

*J*, the described implementation has an efficient memory access pattern.

### Multi-resolution Gaussian basis

The correction term \(x_e\) in model (4) is a tiling of Gaussian functions : \(x_e = G c\), where *c* is the vector of coefficients in the Gaussian basis, and *G* is the operator previously described. In a first approach, all the Gaussian functions (7) have the same variance \(\sigma ^2\), so that operator *G* is linear and problem (10) is convex. The coefficients are placed on a support of size \(N_2^2\) before being (theoretically) convolved with a 2D Gaussian kernel. The spacing between points is *s*, so that the number of required coefficients is approximately \(N_g \simeq \left( \frac{N_2}{s} \right) ^2\).

*G*can be written

*c*: in this case, \(N_g < \left( \frac{N_2}{s} \right) ^2\). In our implementation, the standard deviations are progressively doubled until reaching the diameter of the ROI and then remain constant outside the ROI.

Implementation of (11) is straightforward. The coefficients in vector *c* are classified according to the distance to the center, forming subsets of coefficients \(c^1, c^2, \ldots\). Each subset is point-projected and line-convolved with the corresponding \(\sigma _1, \sigma _2, \ldots\). The resulting sinograms are summed to obtain the projection of *Gc*.

This representation of correction features as Gaussian blobs is actually not a basis in the mathematical sense: some images cannot be represented by a linear combination of Gaussians. However, this representation is very close to a basis for \(\sigma \simeq s\) [20]. In our case, we choose \(s = 0.65 \times \sigma\), meaning that there is a significant overlap between the Gaussians. The discrete Gaussian kernel is truncated at \(3\sigma\), so its length is \(\lceil 6\sigma +1 \rceil\) samples.

### Optimization algorithm

Efficient optimization algorithms can be used to solve the quadratic problem (10). We use the conjugate gradient (CG) algorithm, requiring the computation of the adjoint of the involved operators previously described. CG also entails matrix-vector multiplications, which are efficiently implemented with the CSR representation of point-projector and backprojector.

In the GPU implementation, all the involved arrays are single precision (float 32 bits) as most GPUs are relatively not efficient with 64 bits operations. However, the conjugate gradient algorithm involves scalar products. These operations are implemented by dedicated kernels returning double precision values, as error accumulation is noticeable when accumulating on large arrays in single precision.

## Results and discussion

In the following benchmark, the following notations are used. *N* is the horizontal size of the initial reconstruction, i.e., the diameter of the acquired ROI, which means that the acquired sinogram has a size \(N \times N_p\). \(N_0\) is the horizontal size of the whole object support, unknown in practice (for example, \(N_0 = 512\) in the case of the \(512^2\) Shepp–Logan phantom). \(N_2\) is the horizontal size of the extended reconstruction (\(N_2 > N\)), which should approximate \(N_0\). Lastly, \(N_g\) is the number of Gaussian functions used for the proposed method.

All the tests were performed on a machine with a Intel Xeon CPU E5-2643 12 cores 3.40 GHz and a Nvidia Geforce GTX Titan X GPU. As the LUT can be used for all the slices of a volume, the computation of the LUT is not taken into account. The optimization algorithms used are the preconditioned Chambolle–Pock method [14] for pixel domain exact method and Conjugate Gradient for the proposed method.

Execution time for various local tomography setups

\(N_0\) |
| \(N_2\) | \(N_g\) | Its | Time (s) | PSNR | Its | Time (s) | PSNR |
---|---|---|---|---|---|---|---|---|---|

512 | 272 | 572 | 1345 | 200 | 10.2/2.31 | 35.5 | 4000 | 123 | 36.79 |

512 | 272 | 572 | 729 | 200 | 5.86/2.01 | 34.93 | 3000 | 106 | 35.94 |

1024 | 544 | 1144 | 1345 | 300 | 36.71/5.14 | 28.03 | 4000 | 523 | 31.56 |

1024 | 544 | 1144 | 2081 | 300 | 60.7/11.4 | 30.25 | 8000 | 1094 | 37.85 |

2048 | 1088 | 2288 | 1345 | 500 | 235/33.5 | 27.73 | 4000 | 3570 | 15.13 |

2048 | 1088 | 2288 | 805 | 500 | 129/19.2 | 24.75 | 7000 | 6237 | 20.71 |

4096 | 2176 | 4576 | 2081 | 500 | 1028/109 | 24.11 | 4000 | 14920 (E) | N.A. |

4096 | 2176 | 4576 | 1037 | 500 | 870/97.6 | 22.74 | 7000 | 26110 (E) | N.A. |

Table 1 summarizes the results of the two methods for various setups. For each original phantom size, the two methods are tested with two sets of different parameters. For \(512^2\), \(1024^2\), \(2048^2\), and \(4096^2\) original phantom shapes, the number of projections are, respectively, 800, 1500, 2500, and 4000.

The prototype of [9] was run with the parameters of Table 1. It yields the following execution times: 11.3 s for a \(512^2\) image, 83.1 s for a \(1024^2\) image, 842 s for a \(2048^2\) image, and 3630 s for a \(4096^2\) image. Although it is still better than the “pixel domain approach,” it suffers from very long execution times for large images.

In the example of \(512^2\) phantom size, the proposed method is executed with an acquired sinogram of width 272 pixels. The slice is extended to 572 pixels, and the Gaussian basis is configured to have 1345 functions in total. 200 iterations yield the reconstruction of Fig. 16 in 10.2 s (without taking the LUT computation time). On the other hand, the standard pixel domain method is executed with 4000 iterations and yields a reconstruction similar to Fig. 7, although of slightly lesser quality, in 123 s. The test is then run for a smaller number of Gaussians: the execution time is reduced, but the quality is slightly degraded. This is due to the fact that the number of Gaussians is determined by the spacing *s*, which itself is linked to the standard deviation \(\sigma\). Decreasing the number of unknowns (\(N_g\)) speeds up the computations and also increases the width of the Gaussians, so the reconstruction error might not be appropriately fitted.

The exact method with pixel domain variables starts to be impracticable from \(2048^2\) pixels slices , as thousands of iterations are required to yield an acceptable image quality, leading to hours of processing per slice. The execution times for \(4096^2\) slices were extrapolated from the measured time on 100 iterations: 373 s; therefore, the PSNR are not available in these cases. This method is actually implemented in Python with the ASTRA Toolbox, meaning that only the projection and backprojection are performed on GPU, so the implementation suffers from memory transfers between CPU and GPU. If fully implemented on GPU, one could expect a 5–10 speed-up for this method; nevertheless, the proposed method would still be ahead.

For both methods, the PSNR is progressively decreasing as the size of the slice increases, yet the reconstructions are satisfying. We believe this is a consequence of the cupping being not entirely corrected on the slice borders, which brings more and more contribution as the number of pixels increase.

The current GPU implementation provides acceptable speed-up with respect to the CPU implementation, but there is certainly room for improvement as many parts in the CPU implementation are single-threaded. The computation of the LUT takes several minutes for \(4096^2\) slices, but the same LUT is re-used for all the slices of a volume.

## Conclusion

We proposed a high-performance implementation of a local tomography method aiming at removing the cupping effect. The method consists in iteratively correcting an already reconstructed slice and to reduce the reconstruction error in a Gaussian blobs basis. This implementation is based on a careful analysis of the optimization process, showing that the involved operators can be designed especially for this problem.

Results validate the implementation on simulated data, showing that the known zone constraint effectively enforces an almost zero bias. Benchmarks show that \(4096^2\) slices can be processed in tens of seconds, making it able to cope with modern data volumes.

## Declarations

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

## Authors’ Affiliations

## References

- Herman, G.T.: Fundamentals of Computerized Tomography: Image Reconstruction from Projections. Springer, 233 Spring Street, New York 10013-1578 (2009)Google Scholar
- Clackdoyle, R., Defrise, M.: Tomographic reconstruction in the 21st century. IEEE Signal Process. Mag.
**27**(4), 60–80 (2010)View ArticleGoogle Scholar - Wang, G., Yu, H.: The meaning of interior tomography. Phys. Med. Biol.
**58**(16), 161 (2013)View ArticleGoogle Scholar - Zhao, S., Yang, K., Yang, X.: Reconstruction from truncated projections using mixed extrapolations of exponential and quadratic functions. J. X-ray Sci. Technol.
**19**(2), 155 (2011)Google Scholar - Van Gompel, G.: Towards Accurate Image Reconstruction from Truncated X-ray CT Projections. Ph.D Thesis, University of Antwerp, Antwerp (2009)Google Scholar
- Kudo, H., Courdurier, M., Noo, F., Defrise, M.: Tiny a priori knowledge solves the interior problem in computed tomography. Phys. Med. Biol.
**53**(9), 2207 (2008)View ArticleGoogle Scholar - Yang, J., Yu, H., Jiang, M., Wang, G.: High-order total variation minimization for interior tomography. Inverse Probl.
**26**(3), 035013 (2010)View ArticleGoogle Scholar - Klann, E., Quinto, E.T., Ramlau, R.: Wavelet methods for a weighted sparsity penalty for region of interest tomography. Inverse Probl.
**31**(2), 025001 (2015)View ArticleGoogle Scholar - Paleo, P., Desvignes, M., Mirone, A.: A practical local tomography reconstruction algorithm based on known subregion. J. Synchrotron Radiat. (2016)Google Scholar
- Kyrieleis, A., Titarenko, V., Ibison, M., Connolley, T., Withers, P.: Region-of-interest tomography using filtered backprojection: assessing the practical limits. J. Microsc.
**241**(1), 69–82 (2011)View ArticleGoogle Scholar - van Aarle, W., Palenstijn, W.J., Beenhouwer, J.D., Altantzis, T., Bals, S., Batenburg, K.J., Sijbers, J.: The ASTRA toolbox: a platform for advanced algorithm development in electron tomography. Ultramicroscopy
**157**, 35–47 (2015). doi:https://doi.org/10.1016/j.ultramic.2015.05.002 View ArticleGoogle Scholar - Rashed, E.A., Kudo, H.: Recent advances in interior tomography. Math. Program. 21st Century Algorithms Model.
**1676**, 145–156 (2010)Google Scholar - Vonesch, C., Ramani, S., Unser, M.: Recursive risk estimation for non-linear image deconvolution with a wavelet-domain sparsity constraint. In: 2008 15th IEEE International Conference on Image Processing, pp. 665–668. IEEE, New York (2008)Google Scholar
- Pock, T., Chambolle, A.: Diagonal preconditioning for first order primal-dual algorithms in convex optimization. In: 2011 International Conference on Computer Vision, pp. 1762–1769. IEEE, New York (2011)Google Scholar
- Bilgot, A., Desbat, L., Perrier, V.: Filtered backprojection method and the interior problem in 2D tomography (2009)Google Scholar
- Kannappan, P., Sahoo, P.K.: Rotation invariant separable functions are Gaussian. SIAM J. Math. Anal.
**23**(5), 1342–1351 (1992). doi:https://doi.org/10.1137/0523076 View ArticleGoogle Scholar - Barrett, R., Berry, M.W., Chan, T.E., Demmel, J., Donato, J., Dongarra, J., Eijkhout, V., Pozo, R., Romine, C., van der Vorst, H.: Templates for the solution of linear systems: building blocks for iterative methods vol. 43. SIAM, Philadelphia (1994)View ArticleGoogle Scholar
- Mirone, A., Brun, E., Gouillart, E., Tafforeau, P., Kieffer, J.: The pyhst2 hybrid distributed code for high speed tomographic reconstruction with iterative reconstruction and a priori knowledge capabilities. Nucl. Instrum. Methods Phys. Res. Sect. B Beam Interact. Mater. Atoms
**324**, 41–48 (2014). doi:https://doi.org/10.1016/j.nimb.2013.09.030 View ArticleGoogle Scholar - Niinimäki, K., Siltanen, S., Kolehmainen, V.: Bayesian multiresolution method for local tomography in dental X-ray imaging. Phys. Med. Biol.
**52**(22), 6663 (2007)View ArticleGoogle Scholar - Bale, R.A., Grossman, J.P., Margrave, G.F., Lamoureux, M.P.: Multidimensional partitions of unity and Gaussian terrains. Technical report, CREWES (2002)Google Scholar