A stabilized inversion for limited angle tomography Tim Olsony Department of Mathematics, Bradley Hall Dartmouth College, Hanover, NH 03755 (603) 646 - 3173 Contents 1 Introduction 2 Background and Denitions 2 2 3 Limitations of the Singular Value Decomposition 5 2.1 The Radon transform : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 2 2.2 Tomography and Limited Angle Tomography : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 4 2.3 Physical Motivation and Prior Work : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 4 3.1 Unbounded Inverses and Approximate Identities : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5 3.2 Uncorrelated, Exact Bases and Induced Correlations : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6 3.3 Decreasing Signal To Noise Ratio : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6 4 Mollication Methods 4.1 4.2 4.3 4.4 4.5 4.6 Continuous Approximation to a Discrete Operator : : : Discrete Approximations to the Continous Operator : : Szego's theory for nite Toeplitz operators : : : : : : : : Limited angle spectra : : : : : : : : : : : : : : : : : : : Uncertainty principles and signal recovery : : : : : : : : Non-linear constraints, induced correlations, and POCS : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6 6 7 8 9 10 10 5 The algorithm 6 Implementation 11 12 7 Conclusion 12 6.1 Numerical results : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 12 This work was supported in part y Thanks are due to Dennis Healy, by DARPA as administered by the AFOSR under contract AFOSR-90-0292. Reese Prosser, Dan Rockmore, and Richard Zalik for their advice and encouragement. 1 Abstract In this paper we present a method for reconstructing a function f : R2 ! R from limited angle tomographic data. This reconstruction problem occurs in many physical systems, where physical limitations prohibit the gathering of tomographic data a certain angles. We begin by reviewing some of the classical work on singular value decompositions and POCS in the context of this problem. We then review some of the classical work by G. Szego and others on nite Toeplitz operators. We consider the implications of this work toward a classical inversion of the problem. We then introduce an alternate inversion technique which utilizes multiresolution analysis, induced correlations caused by non-linear constraints, and non-linear ltering to mollify the reconstruction process. We show that the uncertainty principles generated in recent works of Donoho, Stark et al guarantee the invertibility but of this alternative inversion technique. We also utilize the noise reduction techniques of Donoho and Johnstone to reduce the eects of noise on the process. Finally we present numerical examples of the algorithm. We accurately recover images from limited angle tomographic data. 1 Introduction Many problems in applied mathematics involve recovering a function f from measurements of Lf, where L is a known operator. If L is a linear operator, with a bounded inverse, then f can recovered from noisy data Lf + n via standard techniques with little diculty. The recovery of f from noisy data Lf + n is much more dicult if L is an operator whose inverse is unbounded. In this paper we will study the recovery of a function f : R2 ! R, from limited knowledge of its Fourier transform ^f. Thus we are interested in reconstructing f from Lf where L is an operator which reects the limited knowledge of the Fourier transform of f. This problem is motivated by the limited angle tomography problem. This recovery problem is ill-posed and unsolvable without a priori assumptions about f. For example, if f is compactly supported, then f^ will be an analytic function, and therefore f^ will be uniquely determined by its values on any region containing a limit point [3]. Classical analytic continuation techniques are not numerically feasible [25], and the operator L, will generally not have a bounded inverse. A technique for doing analytic continuation was introduced by Papoulis in [22]. This algorithm depends upon the eigenfunctions of the continuous time and band-limiting operators, which were extensively studied in the classical works of Slepian, Pollack, and Landau [33]. The convergence of the algorithm is dependent upon the eigenvalues of the operator, and except in extreme cases, this algorithm is not feasible numerically [10]. Most data recovery problems can be modelled as the inversion of a compact operator L ( for limited data ); where ^ Lf = XS (Xd C f); S is the common support of the functions under consideration, C is the set where we can measure the Fourier transform of f, and XO is a characteristic, or indicator function on the set O. We will to study the inversion of L, from a discrete signal processing perspective. In this context, the continous operator L is naturally viewed as the the limiting case of a series of discrete operators. Since L will generally be compact and self-adjoint, its spectrum can been analyzed via standard techniques. Real-world problems are generated from discretely sampled data sets. Therefore we believe that it more informative to study the spectra of the discrete approximations to L rather than the spectrum of L itself. Our main tool for analyzing the spectra of these discrete approximations to L will be the theory of nite Toeplitz forms, originally introduced by Szego [4]. We will show that the study of these nite Toeplitz forms will allow us to construct an accurate, stable inversion for L, even when the continuous spectra of L suggests that it is not invertible. We will then apply these results to the extensively studied ill-conditioned problem of limited angle tomography. 2 Background and Denitions 2.1 The Radon transform We will review some of the basic properties of the Radon transform. For details and proofs, we recommend Natterer [25]. We let S 1 denote the unit circle, i.e. S 1 = f~ = (cos ; sin )g; 2 where we let denote the angle generated by ~ and the x-axis. We dene a corresponding vector perpendicular to ~ by ~? (? sin ; cos ). Throughout we will use the notation Fn to denote the Fourier transform on L2 (Rn) dened by Z 1 ^ Fn(f)(~s ) f(~s) = (2)(n?1)=2 f(~x)e?i~s~xdx: R We will refer to the band-limiting operators Bl which are most easily dened by their action restricting the Fourier transform of f, i.e. ^ s) ! f(~ ^ s)Xj~sj<l (~s): Bl : f(~ The fundamental theoretical tool behind tomography has always been the Radon transform [31, 25]. The Radon transform, which is sometimes referred to as the x-ray transform or the projection transform, of a function f : R2 ! R is dened by Z (1) Pf (~; ~ ~x) = f(~x + t~? )dt: n R It follows that Pf (~; ~ ~x) is the line integral of f through ~x 2 R2 and perpendicular to ~. We will often abuse this notation by associating the vector ~ with the angle , between ~ and the positive x-axis. An essential property of the Radon transform is that the one-dimensional Fourier transform of Pf (~; ~ ~x) with respect to the variable t ~ ~x is a line through the two-dimensional Fourier transform of f, i.e. ^ ~): F1 Pf (~; ~ ~x) = F2(f)(s~ ) f(s (2) Using this fact and a polar Fourier inversion formula, one can easily relate the Radon transform Pf (~; ~ ~x) to f(~x) through the reconstruction formula f(~x) = Z Z S1 R ^ ~)jsjei(~xs~) dsd = f(s Z Z S1 R F1 Pf (~; ~ ~x) (s)jsjei(~xs~) dsd (3) The inner integral in (3) is simply a one-dimensional inverse Fourier transform. Insertion of an appropriate bandlimiting window into (3) yields the standard ltered backprojection formula [25] fw (~x) = Z Z S1 R F1 Pf (~; ~ ~x) (w(s)jsj)ei(~x~)s dsd = Z S1 Pf (~; ~ ~x) F1?1(w(s)jsj)d: (4) Figure 1: The Radon transform isR illustrated above. The notation ~? stands for the vector which is perpindicular to ~. The line integral Rf (; r) = f(r~ + t~? )dt is taken along the illustrated line at a distance r from the origin. 3 2.2 Tomography and Limited Angle Tomography In this paper, tomography refers to the recovery of a two-dimensional density f : R2 ! R from one-dimensional line integrals, via an inversion formula such as (4). Tomography will generally refer to the recovery of such a two dimensional density from a full complement of one-dimensional line integrals, i.e. Pf (~; ~ ~x) for all ~ 2 S 1 , or an evenly subsampled version of these line integrals. Typical sampling algorithms conditions can be found in Natterer [25]. Limited angle tomography refers to the recovery of a function f from a proper subset of its line integrals. More specically, in limited angle tomography only the line integrals Pf (~; ~ ~x) where jj < < =2 are available for the inversion. From (2) we know that the one-dimensional Fourier transform of each projection corresponds to a line through the origin of the two dimensional Fourier transform of the original image. The diculty of recovering a function from limited angle tomographic data can be understood through the reconstruction formulas (2),(3), and (4). We can see via (2) that if the projection data Pf (~; ~ ~x) is only available ^ ~) will only for a restricted angular range, ? < < , where < =2 then the Fourier data F2(f)(s~) f(s be available for the restricted angular range ? < < . We know from (3), and (4) that all of the Fourier, or projection data is needed to accurately reconstruct f. Thus we are forced to try to recover an unknown portion of the Fourier transform of f, as is illustrated in Figure 2. Known Fourier Data Unknown Unknown Fourier Data Fourier Data Known Fourier Data Figure 2: The limited data problem generated by limited angle tomography is illustrated above. In order to accurately reconstruct the original density, we must recover an unknown portion of the Fourier transform of f. If we know nothing about the objective function f : R2 ! R, then this reconstruction would certainly be impossible. Usually, though, the support of the object is nite and known. It is also generally the case that the function is positive and real. Therefore, we assume throughout that we are interested in reconstructing a bounded function f : S \ R2 ! R, from the values of its Fourier transform, where S is a compact set, containing the origin. This will imply that the Fourier transform of f will be an entire function of nite exponential type, and therefore the continuous problem is solvable. Classical analytic continuation methods, however, are not numerically feasible [10, 25]. The creation of a stable numerical inversion for this problem is the subject of this paper. 2.3 Physical Motivation and Prior Work The limited angle tomography problem arises in a number of contexts. In biology, transmission electron microscopy is used to image the structure of chromatin bers in tissue. Because of the rapid absorbtion of the electrons an extremely thin slice of tissue is prepared. This slice becomes prohibitively thick when the angle of incidence of the rays goes beyond a certain threshold. Therefore only projections at angles within approximately 60 degrees of the normal to the slice can be utilized. Another application occurs in the non-destructive testing of welds in critical structures such as nuclear reactors, rockets, and missiles. In this case, as before, data can only be collected at angles which do not cause severe attenuation of the beams. From these examples we are able to generalize and see that there are a great number of applications in which an accurate limited angle tomography algorithm would be extremely useful. 4 Limited angle tomography has been extensively studied. It has been proven [7, 9, 16] that when the singular values of the associated operator are viewed as a function of the angular limit , they converge to zero extremely quickly and as a result the problem is highly ill-conditioned. One approach to the problem of reconstructing a density function from limited angle data is to compute the projection of the known data onto the the range of the Radon transform [25]. The result is a completed data set, and this data set is then processed via typical reconstruction algorithms to produce an approximation to the original density function. The range theorems for the Radon transform make it possible to recover the low frequency components of the image through this process. The crucial step in this process is the determination of the coecients of a trigonometric polynomial, from non-uniform samples. The high frequency components of the image require the recovery of high order trigonometric polynomials, however. This process is notoriously ill-conditioned, and as a result the algorithm will not perform well when the angular limit is not close to =2 and a high resolution reconstruction is desired. In [32] a alternative reconstruction algorithm was proposed. It was speculated that it might be close to optimal. It was shown in [34], however, that this algorithm formed a commutative diagram with ltered backprojection, and as a result was of little use in limited angle tomography. A recent paper by Quinto [30], uses the microlocal analysis concept of wavefront sets to quantify the information which would be dicult to recover in limited data tomography. This work reinforces with the general feeling in the area that the unknown high frequency information, which is generated by singularites, will be very dicult to recover. There has been recent work by Ramm and Zaslovsky [26], which proposed the recovery of singularities of a two-dimensional density from limited angle tomographic data using the Legendre transform. There were also theoretical stability results. We are not aware of any numerical application of this result, however. Finally, a thorough review of other approaches to the limited angle tomography problem can be found in Natterer [25]. 3 Limitations of the Singular Value Decomposition 3.1 Unbounded Inverses and Approximate Identities The standard numerical method for inverting a linear operator is the singular value decomposition (SVD). The singular value decomposition of a linear operator L consists of the normalized eigenfunctions of L L, f n g, and the p functions fng = f1= n L ng wherepn is the eigenvalue corresponding to n . The vectors n and p n are the singular vectors, and the values n = n are the singular values. It follows that both f n g and f1= n A ng are orthonormal sets for the domain and range of L. Most important is the fact that these vectors allow a decomposition of L given by X Lf = nhf; n in: (5) n Moreover, the singular value decomposition also yields an inversion formula f= X ?1 n hLf; n i n : n (6) A great number of problems in applied mathematics can be reduced to the inversion of a linear operator, and as a result, the singular value decomposition has been extremely important in applied and numerical mathematics. Unfortunately, since the operator L is usually compact, the values n ! 0 as n ! 1 [1]. Thus when only nite precision is available in both the measurements and computer facilities, the usefulness of (6) is somewhat limited for two reasons. The rst is that the storage of the data Lf with nite precision will cause the higher order terms in the sum (5) to be beyond the wordsize of the machine, and thus they will be rounded to zero. In addition, the multipliers n?1 in (6) will approach 1 as n ! 1. This makes the computation of the higher order terms in the sum in (6) very unstable or impossible and therefore an altered, truncated version of (6) is often used. These mollied methods, such as the truncated SVD, and Tikhonov-Phillips [25] regularization can usually be represented in the form X Mf = w(n)n?1 hLf; ni n n where the windowing function w(n) is chosen so that w(n)n?1 2 l2. For instance the truncated SVD would dene w(n) = 1 for n < N and zero for n N. Thus an approximate inversion is obtained, which is computationally stable, but not exact. 5 The second problem with the SVD is that if there are two distinct classes of singular values: those which are close to a xed value b > 0 and those which are nearly 0, then the inversion formulas will all reduce to an approximation of a truncated SVD. If in addition, the operator is self-adjoint, the functions f ng will be equal to the functions fng, and will be eigenvectors. In this case the truncated SVD will act as an approximate identity, since division by the singular values which are all in the vicinity of b will only scale the function Lf, and the singular values which are nearly 0 will truncated from the expansion. Thus the eort of constructing and inverting the SVD will not cause a noticeable improvement in the nal result. As we will show in Section 4.3, this is the case in limited angle tomography. The operator is compact and selfadjoint, and the singular values, or eigenvalues, of the operator separate distinctly into two classes, those which are nearly 2 and those which are nearly 0. 3.2 Uncorrelated, Exact Bases and Induced Correlations The SVD produces exact orthonormal bases for the range and domain of any compact linear operator. Although there are many instances when an exact representation is useful, an exact representation is not desireable in this type of situation. When reconstruction problems, or inversion problems are ill-conditioned, we believe that an overcomplete basis, which is correlated and redundant, is more useful than an exact basis. It is well known that in limited angle tomography, and many other imaging problems, low resolution information can easily be reconstructed. With an exact, uncorrelated basis, this low resolution information will not help to reconstruct the high frequency details which are dicult to recover. If a correlated, redundant basis, is used, however, then the low resolution information can aid in recovering the high frequency details. We will utilize non-linear or linear constraints to induce correlations between the coecients of an orthonormal basis. In our case, the initial non-linear constraints are positivity, compact support, and the known data. Once we reconstruct the low-resolution information in the image, we will then use this low-resolution information as an additional non-linear constraint. 3.3 Decreasing Signal To Noise Ratio The linear POCS, algorithm, as outlined in [22], uses the singular functions of the time- and band-limitingoperators to construct an approximate inverse to the problem. More specically, if f is the desired function, with the eigenfunction P expansion for f = cnen , and we observe a blurred version of f, namely Lf, then the K th step of POCS yields the result X P K (Lf) = (1 ? (1 ? n)K )cnen : We observe that as K ! 1 this converges to f. P In the presence of white noise, however, we will observe Lf = (n cn + gn)en , where gn are i.i.d. observations of white noise. Now the reconstruction procedure on the K th step will yield X P K (f + g) = (1 ? (1 ? n )K )(cn + 1 gn)en : n As K ! 1, it follows that the signal to noise ration of the solution will go to zero ( assuming that L was an innite dimensional compact operator ). Moreover, the solution will converge to X (cn + 1 gn)en : n Since it is usually the case that n ! 0, this solution is not desireable. 4 Mollication Methods 4.1 Continuous Approximation to a Discrete Operator A singular value decomposition for the operator was constructed in [9], and the asymptotics of the singular values of the operator were investigated. It was proven that these singular values, when viewed as a function of the angular limit , converge to zero extremely quickly. The singular value decomposition was also studied in [16]. There it was shown that the condition number grows only linearly, for near =2, but grows much more quickly for smaller . These studies concentrated on the continuous operator, rather than its discrete realization, however. 6 We will study the behavior of the spectra of the discrete approximations to the continuous operator L, for xed , as a function of the discretization. We will show that the an appropriate Hilbert-Schmidt norm of these discretizations only grows linearly, and therefore, these discretized operators are not horribly ill-conditioned. It was noted by Davison in [9] that \Our results should not be construed as proof that limited angle tomography is impossible". Unfortunately, those results and the work of others on the SVD for the limited angle problem, have led some to believe that limited angle tomography is impossible. It was also noted in [9] that \In practice there are often important a priori constraints on the unknown function f", and that these constraints \have no natural expression in terms of the singular functions." We will incorporate these constraints into our algorithm, in order to help stabilize the problem. We will present a method which uses a series of singular value decompositions, at increasingly ne resolutions, to construct a stabilized inversion of the limited angle tomography problem. This stabilized inversion signicantly outperforms the singular value decomposition for the problem. The importance of this type of approach is partially due to the fact that it concentrates on the actual discrete inversion of the operator in question, rather than a continuous approximation to the discrete inversion, or vice versa. In this context, the continuous operator is naturally seen as a limiting case of the discrete operators. Moreover, the spectrum of the continuous operator is necessarily worse than its discrete approximations, and thus we might be led to believe that the problem is intractable, when in fact it is not as bad as it might seem. 4.2 Discrete Approximations to the Continous Operator We will now outline the conversion of the limited angle tomography problem into a discrete signal processing problem. The reconstruction process will begin with the collection of a nite number of measurements of the Radon transform. A fundamental principle of signal processing is that innite frequency resolution is not possible, given only nite measurements. In addition, the nal outcome of any imaging process is a N N pixel array of values which are displayed for the scientist, radiologist, etc. This N N array should represent the properly sampled values of the function f. Since there were can only be a nite number of input measurements, and the output is a nite array, it follows that only a nite number of frequencies are present in the reconstruction, (although some may be due to aliasing.). ^ [?; ] [?; ]. This is not inconsistent Therefore we assume that f is a band-limited function, i.e. supp(f) with our assumption of compact support in the former section. Rather, we should think of f as being the image of a compactly supported function, under the band-limiting operator. We also need an assumption concerning the essential support of the function f in time. We will be concerned with the recovery of functions f 2 L2 (R2) which are band-limited, and are essentially compactly supported in the following ^ = [?; ] [?; ], way. Since we have assumed that the functions under consideration are band-limited, i.e. supp(f) we can represent them as sums of sinc functions, i.e. XX (x ? m) sin (y ? n) : f(x; y) = f(m; n) sin(x ? m) (y ? n) We assume that all of the functions f are essentially compactly supported in the sense that they have nite sinc expansions: f(x; y) = N X N X 1 1 (x ? m) sin (y ? n) : f(m; n) sin(x ? m) (y ? n) We will refer to the size of the pixel array, N, for an N N as above the discretization size of the operator. These assumptions are consistent with the reconstruction of a compactly supported object. If we restrict ourselves to the case where we know only 90 degrees of projection, or equivalently Fourier data as is illustrated in Figure 3. If we assume that the rst and third quadrant of the Fourier transform is given then the operator assumes the form Lf = XS N X N X ?N ?N 2 (x ? m) sin 2 (y ? n) cos ((x ? m) + (y ? n)): f(m; n) sin (x ? m) (y ? n) 2 2 2 (7) This formula can be most easily realized by expanding the Fourier transform of f in an exponential series on [?; ] [?; ], and seeing that the terms in the sum are translated versions of the inverse Fourier transform of the characteristic function on the known data in Figure 3. 7 Unknown Known Fourier Fourier Data Data Known Unknown Fourier Fourier Data Data Figure 3: We illustrate the known data in a 90 degree Fourier reconstruction problem. Since we are only interested in the values attained by Lf at the integers, which will represent our pixel values, this naturally generates a linear operator from the original set of sample values f(m; n) to the new values Lf(m; n). This operator can most eciently be viewed as a truncated discrete convolution of 2 (m) sin 2 (n) cos (m + n) K(m; n) = sin (m) (n) 2 2 2 sampled at the integers against the array ff(m; n)g. We will now study the behavior of nite convolution, or Toeplitz operators of this type. 4.3 Szego's theory for nite Toeplitz operators In this section we will study the spectral properties of discrete Toeplitz operators generated from compact operators of the type Lf = XS F2?1(f^XC ): We will study the spectral properties of these discrete operators, as a function of the discretization N. We will then use this study to introduce an altered inversion formula which will prove to be stable. Szego's one-dimensional theory of Toeplitz operators extends, essentially with a change of notation, to ndimensions. For completeness, we will present a small portion of this here. What follows can be found in the one-dimensional case in [5]. We begin with a convolution kernel K(m; n) which is a symmetric, square summable function in l2 (Z2), and ^ It follows that K^ is an integrable real function, and for denote its Fourier transform on [?; ] [?; ] by K. notational purposes we dene Z 1 ^ s)ei~s ds K() = c = j j K(~ ^ or the values of K, and 2 Z Z. to be the Fourier coecients of the K, Let SN be a subset of the lattice of integers Z2 R2, which represents the eective support of our function, i.e. SN = [0; 1; 2; :::N ? 1] [0; 1; 2; :::N ? 1] = ZN ZN where N is our discretization. The operator (7) can then be viewed as the nite Toeplitz, or convolution operator TN (K)(s) = X 2 SN c ? s ; where K^ = XC . We will assume throughout that XC is symmetric about the origin, and it then follows that TN (K) is a symmetric operator, and thus has real eigenvalues. If we let v be an eigenvector, with unit norm and eigenvalue , it follows that X X (8) = TN (K)(v) v = c ?v v : 2 2 S N SN 8 Utilizing the denition of c along with a basic combinatorial identity we can express (8) in the following form: Z X X X ^ = c ? v v = j 1 j j v ei sj2K(s)ds: (9) 2S 2S 2S N N N Assuming that m ess inf K < ess sup K M, it then follows from basic measure theory arguments that m < < M: In our case, we had normalized the kernel so that K(~0) = 1, and thus the Fourier transform of the kernel was K^ = 2XC , where C is the set where we know the Fourier transform. Thus 0 = m < < M = 2 for each eigenvalue and we can now see that our operators are non-singular, for each discretization N. Much more is known about the eigenvalues of Toeplitz forms [5]. As the discretization N ! 1, the eigenvalues become distributionally equivalent to K^ in the following sense. The distribution of the eigenvalues fN g2S of nite Toeplitz operators, converges to the distribution of the values of K(~s), i.e. N lim N !1 jfN ja < < bgj 1 ^ s) < bgj; = j j jf~s ja < K(~ jSN j as long as the measure of the sets where F(~s) = a or F(~s) = b is zero. In words, the proportion of eigenvalues in a given interval, converges to the measure of the proportion of the set where F is in the corresponding interval. To gain some insight into the ill-conditioned nature of inverting any limited angle problem, we will apply (10) to the nite Toeplitz forms associate with the Fourier transform of our kernel K^ = 2XC . Since all of the eigenvalues are positive, it follows that jfN j0 < < gj jfN j ? < < gj 1 ^ s) < gj = j j ? jC j : lim = lim = j j jf~s j ? < K(~ (10) N !1 N !1 jSN j jSN j j j Thus the proportion of eigenvalues which eventually tend toward 0 is exactly the same as the proportion of where K is 0, or the measure of n C. More explicitly, lim N !1 jfN j0 < < gj = jUnknown Dataj : jSN j jNecessary Dataj (11) In the case which we have been studying, i.e. the 90 degree case, half of the eigenvectors must approach 0, and half must approach 2. This discussion highlights the problems associated with using a truncated or mollied SVD to invert this type of operator, when the discretization N is large. There will be eigenvalues which are essentially zero, which will represent the unknown data, and eigenvalues which are near 2, which represent the known data. If N is large, the truncated SVD will only rescale the nal image, by dividing the known data by 2. The unknown coecients will be truncated to zero, and no new information will be gathered. 4.4 Limited angle spectra We have noted above that the spectra of limited data operators separate into two classes. There are eigenvalues which are essentially one, and eigenvalues which are essentially zero. Because of the above discussion of the geometry of the eigenvectors, it is not at all surprising that half of the eigenvalues are greater than 1=2, and nearly 1, and half of the eigenvalues are less than 1=2, and nearly 0. For the 90 degree limited angle problem, 8944 out of 16384 eigenvalues the are greater than .01, or 54%. Thus a truncated singular value decomposition with a tolerance of .01 will recover only 4% more information than was present in the original 50% of the known data. The dierence in image quality due to this recovered information is negligible. Thus a truncated singular value decomposition, used at at discretization of 128x128 will act essentially as an identity operator on the blurred image, rather than an eective inversion tool. We now consider the spectra of the limited angle operators from smaller discretizations. In Figure 4, we have plotted the spectra of the limited angle operators at various discretizations (note that normalizing factors cause them to tend toward either 0 or 2. From the gure, we see that multiresolution analysis can be used to mollify of the inversion. The low resolution operators are not massively ill-conditioned, and we should be able to reverse the eect of the blurring at low resolutions. We want to use this information to help reverse the blurring at higher resolutions. 9 4.5 Uncertainty principles and signal recovery Another way to understand the diculties of signal recovery, and the advantages of our approach, is via the work of Donoho and Stark on uncertainty principles [10]. The classical uncertainty principle states that if a function f(t) is ^ is essentially zero outside an interval essentially zero outside and interval of length t and its Fourier transform f(w) of length w, then tw 1: A basic tenet of [10] is the following. Suppose that you are trying to recover a signal f(t) which are non-zero on ^ Suppose further that for some reason you do not a region of size t, from knowledge of the Fourier transform f(!). ^ know the values of f(!) on some interval I where 1: jI j < w t ^ can not be essentially zero outside of I, and thus you can observe f(t) from Then by the uncertainty principle f(!) ^ where ! 62 I. Since this problem is linear, this is equivalent to stating that the problem is invertable if and only f(!) if the kernel of the operator is 0. Donoho and Stark constructed discrete analogues of the uncertainty principle which are of use here. The rst of these is the following Lemma 1 Let fxtg be a discrete sequence with at most Nt non-zero elements, and let fx^!g be its discrete Fourier transform. Then fx^! g cannot have Nt consecutive zeros. This Lemma can be generalized to an n?dimensional discrete Fourier transform, with a proof essentially identical to that in [10], except that the Vandermonde matrix of [10] will be replaced with a tensor product of Vandermonde matrices. We state this below. Lemma 2 Let xt be a two dimensional sequence on S = ZN ZN, and let x^! be its two dimensional discrete Fourier transform. If xt is non-zero only on a block of the type R = [r1; r1 + 1; :::r1 + n] [r2; r2 + 1; :::r2 + n] S then x^! cannot vanish on a block Q = [q1; q1 + 1; :::q1 + n] [q2; q2 + 1; :::q2 + n] S . We use this to construct the following two-dimensional result, which applies directly to our problem. Suppose that a function ft , where t 2 S is non-zero only on the interior N=2N=2 region of S. Then we can recover it from the values of its discrete Fourier transform on the region C = [0; 1; ::::N=2 ? 1] [0; 1; ::::N=2 ? 1] [ [N=2; N=2+ 1; ::::N ? 1] [N=2; N=2+1; ::::N ? 1]. This result follows directly from the above Lemma, since the problem is linear ^ cannot vanish on either [0; 1; :::N=2 ? 1] [N=2;N=2+1; ::::N ? 1] or [N=2; N=2+1; ::::N ? 1] [0; 1;:::N=2 ? 1], and f(!) thus the problem must be non-singular. This is of interest in our case, because the set C is the discrete analogue of the rst and third quadrants in the continuous case. Thus we nd that if we oversample the discrete Fourier transform of a function ft by a factor of 2, the recovery of the original function from limited angle Fourier data is non-singular. This corresponds with having a \true" rather than a modulo convolution on the support of the image. The problem is very ill-conditioned, however, if the discretization is large. We once again return to the \moral" argument that a signal recovery problem is invertible, if the data to be retrieved cannot vanish on the unknown region. Although the data cannot vanish on the unknown region, it can essentially vanish, if the discretization is large. Thus, if we try to recover ne details from the beginning, the problem will be very ill conditioned. If however, we can recover a low-resolution reconstruction rst, and use this low resolution reconstraction to mollify the ne-resolution reconstructions, then we might have some hope. 4.6 Non-linear constraints, induced correlations, and POCS We assume throughout that the functions which we are recovering are compactly supported, and have nite support, which is known to some degree. Moreover, we know a portion of the Fourier transform of these functions. Finally, we assume that there is some maximum density which can be observed, i.e. gold markers in electron microscopy. Thus our algorithm begins with the constraints on our image f: 1. 0 f M. 2. f^XC is known. 3. suppf 10 The techniques which we will use to stabilize the algorithm are highlighted below. Multiresolution analysis: It is well known that a low resolution version of f can be reconstructed accurately. This low resolution version of f is then added to the constraints. If we utilize a projection operators Pk for the projections onto these resolution subspaces, then it follows that we are interested in using our ability to recover P1f, or a low resolution version of f, to aid us in the recovery of f on higher subspaces. Preservation of non-linear constraints: We would like the non-linear constraints outlined earlier to hold on these subspaces. In other words, we would like for Pnf to be positive, and compactly supported. This is the case if we use the Haar functions for our reconstruction. Induced correlations: We want to preserve the non-linear constraints on the low-resolution subspaces, so that the dierence subspaces Dn = Pn n Pn?1 will be correlated to Pn?1. In the case of the Haar subspaces, positivity implies that if the scaling coecient at level n ? 1 is given by sk;n?1, then the dierence coecients on this interval will be bounded bounded above and below by s : ?1 Similarly, the support constraint induces correlations between coecients at dierent scales. This is of utmost importance to this algorithm. If we think only of the linear problem, then one would have to retrieve cnen when one has only observed the noisy coecient (n cn + gn )en . If n is small, this is a very ill-conditioned problem. But if there is a great deal of correlation between scales, then it implies a great deal of correlation between singular functions. Thus we can observe cn en from the coecients ck ek where k < n. The \low resolution" information will allow us to recover the \high resolution information". Signal to Noise Recovery As outlined previously, the linear POCS algorithm will necessarily produce a very noisy image, if the input data is noisy. More specically, the signal to noise ratio's will be essentially driven to zero as n ! 1 (although this assumes the existence of truly white noise, which would require innite noise power, the approximation in nite dimensions is essentially true). To combat this we want to denoise the reconstruction periodically, in order to restore the dominant features and remove the noise. We do this with the non-linear thresholding techniques of Donoho and Johnstone [11, 12]. We utilize this type of technique because images are generally dominated by their edges. Non-linear thresholding is designed, to allow edges to remain sharp, while polynomially smooth regions can be smoothed or denoised. Since this is incorporated into the more general POCS algorithm, we will use the \hard" non-linear thresholding algorithm, rather than the slightly more stylish \soft" non-linear thresholding. k;n 5 The algorithm Our algorithm consists of three dierent procedures. Stable low-resolution reconstruction We have shown that when the resolution of the problem, or the size of the numerical grid SN is suciently large, the spectrum of the problem will make a truncated or mollied SVD essentially useless. Thus we will not be able to eectively invert an operator of the type Lf = XB F2?1(f^XC ) unless the discretization is suciently coarse, or the frequency resolution is very low. If the discretization is suciently coarse, however, the spectra of the discrete operators is quite dierent, as can be see in Figure 4. Therefore we will begin by recovering the image on a very low-resolution Haar subspace, i.e. we begin by recovering local averages f1. This will serve as the rst approximation to the desired object function f. We reiterate that f1 will be the image of f in a low-resolution Haar subspace. Stable iterative improvement: Once we have a rst approximation f1 to f, the natural question to ask is whether we can use f1 and our original data Lf to reconstruct a higher resolution reconstruction of f, which we will call f2 . As outlined before, the Haar subspaces preserve the non-linear constraint of positivity, as well as the linear constraint of compact support. These constraints induce correlations between the scales, and allow the low resolution subspace to mollify the reconstruction on the higher resolution subspace. Thus we use the induced non-linear constraint given by the solution on the former subspace, as well as positivity, which induces a bound on the size of the coecients in the dierence subspace. We also utilize the compact support for further stabilization. 11 Iterative error reduction: As outlined before, any type of algorithm which yields an actual inverse to a noisy ill-conditioned problem, will drive the signal to noise ratio in the nal image to zero. Therefore we need to remove noise from the image along the way. We utilize both (POCS), and non-linear shrinkage to accomplish this goal. Projection onto convex sets, or (POCS) [23, 2], is a standard iterative algorithm for utilizing a priori knowledge in a reconstruction algorithm. We assume throughout that 0 f M, supp f BN , and that we know f^XC . All of these assumptions are consistent with tomography, where positive, compactly supported densities are recovered. Moreover, the maximum density of the image is also well known, (i.e. bone when doing medical imaging, gold tracers when doing electron microscopy, etc.). Thus we have three convex sets, corresponding to these three assumptions, and we want our answer to lie in the intersection of these convex sets. This technique allows us to reduce the residual computational error during the algorithm, and stabilize the nal outcome. 6 Implementation We will now explicitly outline the algorithm. 1. Using POCS, with the constraints outlined above, on the input data. After a short while you will discover that you have a good low resolution Haar approximation to your function. 2. After attaining a good low resolution approximation, use the low resolution approximation as an additional constraint, and once again, allow POCS to run, with the additional constraint. 3. As each successive approximation converges on each subspace, reduce the noise in the image via non-linear \hard" shinkage as in [11, 12]. 4. Iterate steps 2 and 3, at higher and higher resolutions. 6.1 Numerical results We will illustrate our algorithm with three separate numerical experiments. We do not feel that these experiments are all inclusive, but we do feel that they demonstrate the potential of this type of algorithm. All of the experiments begin with limited angle projection data, and use the second Fourier reconstruction algorithm, outlined in Natterer [25], to construct the \known" portion of the Fourier data. The algorithm is then run as is described above. In all experiments, the known \full" data set consists of 256 evenly spaced angles, and 256 samples per angle. The limited angle data sets consist of correspondingly less angles. In the rst experiment, we use 145 degrees of projection data. The only \noise" which the algorithm will be subject to the is \noise" created by the Fourier reconstruction algorithm. In this rst experiment, we assumed that knowledge concerning the support of the object is not exact. We assumed only that the support of the object is embedded in a circle inscribed the reconstruction square. Nevertheless, the algorithm recovered the edges correctly, although their magnitudes were not exactly reconstructed. The results of this experiment can be seen in Figure 6 Note that there is one irritating portion of the reconstruction, namely the widened edge at the top right. In the second experiment, we use 145 degrees of noisy projection data. We also assumed that we knew the support of the object fairly exactly. The results of this experiment can be seen in Figure 7. The third experiment utilizes only 90 degrees of projection data, and once again assumes knowledge of the support of the object. The reconstruction can be seen in Figure 8. We should note at this time, that it seems that utilizing the complete knowledge of the support of the object, is useful, but also can cause artifacts. If the data is not exact, a full data reconstruction will not produce an object with exactly the support of the original image. As a result, insisting on this constraint can cause artifacts. This constraint seemed necessary, however, if one was to recover the image from the extreme limited angle case of 90 degrees. 7 Conclusion Upon examining the procedure we get the feeling that we are getting something for nothing. The explanation for why we are able to extrapolate into the missing region of the Fourier transform is simple. In order to get the true 12 convolution of the image against our kernel, we have to zero pad, and therefore oversample the Fourier transform of the function f(x; y). Since f has compact support this is equivalent to interpolating with a sinc series which is very non-local. Therefore we are implicitly getting non-local information about our Fourier transform. This is the information which we are utilizing in order to recover the missing region of the Fourier transform. This conforms with the constraints of the limited angle tomography problem, because it is equivalent to zero padding the projections, at each angle, or oversampling the Fourier transform of the projections. There are a number of issues which we have not addressed. We have utilized empirical timing schedules to decide when to impose the low resolution constraints on our algorithm. We believe, however, that these timing schedules can either be replaced with xed schedules, or with adaptive schedules as in simulated annealing. We have also utilized empirical results to ne tune the non-linear ltering. The non-linear ltering has been extensively studied and this does not seem to be a concern. A recent paper by Quinto, [30], states that \if a singularity is not stably visible from limited data, no algorithm can reconstruct it stably". Our results support the fact that these are the dicult edges to recover, but they are quite a bit more optimistic. The results of [30] are generated from Sobolev norms and microlocal analysis, and are very elegant. In practical numerical analysis, however, decay at innity is not really an issue. It is well known that the low frequencies can be reconstructed [25]. We use the low frequencies to build outward and recover edge detail. In a certain sense, we are solving a series of problems which are much easier than the limited angle tomography problem. We certainly agree, however, that the edges which are tangent to the view directions will be much easier to reconstruct. Another issue which should be addressed is that in most applications where limited angle tomography is needed, the object is not eectively compactly supported. The wall of a nuclear reactor is better modeled as innitely long for all practical purposes. A weld is generally a local non-uniformity in a uniform substance, however, so the projections can be changed synthetically, after they are gathered, in order to make them correspond to the local, compactly supported region of interest. In electron microscopy, however, the solution to this problem is not obvious. References [1] M.Reed and B.Simon, \Functional Analysis I", Acad. Press, 1980. [2] D.C. Youla, \Mathematical Theory of Image Restoration by the Method of Convex Projections", in Image Recovery, Theory and Applications, AP, San Diego, 1987. [3] I.M. Gelfand and G.E. Silov, \Fourier transforms of rapidly increasing functions and questions of the uniqueness of the solution of Cauchy's Problem.", Am. Math. Soc. Trans., Vol. 15, pp. 221-283. [4] G.Szego, \Ein Grenzwertsatz uber die Toeplitzschen Determinanten einer reellen positiven Funktion", Math.Ann. 76, 490-503, 1915. [5] U.Grenander and G.Szego, Toeplitz Forms and Their Applications, Univ. of Cal. Press, 1958. [6] I.I.Hirshman, Jr. and D.E. Hughes, \Extreme Eigen Values of Toeplitz Operators", Lecture Notes in Mathematics, No. 618, Springer-Verlag, 1977. [7] F.A.Grunbaum, \A study of Fourier space methods for \Limited Angle" image reconstruction," Num Func Anal Opt, (1980), 2(1), 31-42. [8] M.E.Davison and F.A.Grubaum, \Tomographic reconstructions with arbitrary directions", Comm. Pure and Appl. Math, 34, (1981), 77-120. [9] M.E.Davison, \The ill-conditioned nature of the limit angle tomography problem," SIAM J.Appl.Math, 43, April 1983, 428-448. [10] D.L. Donoho and P.B. Stark, \Uncertainty Principles and Signal Recovery", SIAM J. Appl Math, Vol. 49, No.3, pp. 906-931, June, 1989. [11] D.L. Donoho and I.M. Johnstone, \Minimax estimation via wavelet shrinkage", Tech. Rep. 402, Dept. of Stats., Stanford University, Stanford, CA. [12] D.L. Donoho and I.M. Johnstone, \Ideal spatial adaptation by wavelet shrinkage" ", Tech. Rep. 400, Dept. of Stats., Stanford University, Stanford, CA. 13 [13] D. Jackson, \U ber die Genauigkeit der Annaherung Stetiger Funtionen Durch Gegebener Ordnung", dissertation, Gottingen, 1911. [14] W.D.Cheney, Introduction to Approximation Theory, Chelsea, 1966. [15] A.K.Louis, \Picture reconstruction from projections in restricted range", Math. Meth. in the Appl.Sci.,2 (1980), 209-220. [16] A.K. Louis, \Incomplete data problems in x-ray computerized tomography I", Num.Mathematik, 48, (1986), 251-262. [17] A.K. Louis and A. Reider, \Incomplete Data Problems in X-Ray Computerized Tomography II", Num.Mathematik, 56, (1989), 371-383. [18] P.Maass, \The X-ray transform: singular value decomposition and resolution", Inverse Problems, 3 (1987), 729-741 [19] W.R.Madych, Summability and Approximate Reconstruction from Radon Transform Data, Contemporary Mathematics, 113, pp. 189-219, 1990. [20] W.R.Madych and S.A.Nelson, \Reconstruction from restricted Radon transform data: resolution and illconditionedness", SIAM J.Math.Anal, 17, 1447-1453. [21] S.Mallat, \Multiresolution representation and wavelet orthonormal bases of L2 (R)", Trans.Amer.Math.Soc., 315 (1989), 69-87. [22] A.Papoulis, \A New Algorithm in Spectral Analysis and Band-Limited Extrapolation", IEEE Trans on Cir and Sys, 9, Sept. 1975. [23] M.I.Sezan and Henry Stark, \Applications of Convex Projections Theory to Image Recovery in tomography and Related Areas", Image Recovery: Theory and Application, Acad Press 1987, Henry Stark, Ed. [24] R.A.Horn and C.R.Johnson, Matrix Analysis, Cambridge, 1990. [25] F.Natterer, The Mathematics of Computerized Tomography, John Wiley, 1986. [26] A.G.Rammand A.I.Zaslavsky, \Reconstructing singularities of a function from its Radon transform,", preprint. [27] A.G.Ramm, \Inversion of limited-angle tomographic data", Comp. and Math. with Appl., 22, 1991, 101-112. [28] E.T.Quinto, \Singular value decompositions and inversion methods for the exterior Radon transform and a spherical transform", J.Math.Anal.Appl., 95 (1983), 437-448. [29] E.T.Quinto, \Tomographic reconstruction from incomplete data- numerical inversion of the exterior Radon transform", Inverse Problems, 4, (1988), 867-876. [30] E.T.Quinto, \Singularities of the x-ray transform and limited data tomography in R2 and R3.", SIAM J.Math.Anal. to appear, 1993. [31] J.Radon, \U ber die Bestimmung von Funktionen durch ihre Integralwerte langs gewisser Mannigfaltigkeiten.", Berichte Sachsishe Akademie der Wissenschaften, Liepzig, Math.-Phys.Kl, 69,262-267. [32] J.A. Reeds and L.A. Shepp, \Limited Angle Reconstruction in Tomography via Squashing", Trans. on Med. Im., Vol. MI-6, No. 2, June, 1987, 89-97. [33] D. Slepian, and H.O. Pollack, \Prolate spheroidal wave functions, Fourier analysis and uncertainty, I and II", Bell Syst. Tech. J., 40, (1961), pp. 43-84. [34] T. Olson and J. Jae, \An explanation of the eects of Squashing in limited angle tomography", Trans. on Med. Im., Vol. 9, No. 3, Sept. 1990. 14 N=6 N=8 2 2 1 1 0 5 10 0 15 10 N=12 2 1 1 20 40 0 60 40 N=24 2 1 1 80 160 0 240 160 N=48 2 1 1 320 640 120 320 480 N=64 2 0 80 N=32 2 0 30 N=16 2 0 20 0 960 600 1200 1800 Figure 4: Above we illustrate the spectra of the 90 degree limited angle operators. Notice that as the discretization becomes large, the eigenvalues converge to either 0 or 1. 15 (a) (b) Figure 5: The discrete invertability result is illustrated above. In (a), we illustrate the support of f, on the discrete lattice SN = ZN ZN . In (b), the known portions of the discrete Fourier transform f,^ are illustrated. These known regions correspond to the known regions of the Fourier transform in limited angle tomography. The discrete invertability result implies that f^ cannot vanish on these known regions. Figure 6: Above we illustrate the reconstruction of the Shepp-Logan phantom from 145 degrees of data. The full data reconstruction is shown at top left. The reconstruction using only the limited data is shown at top right. The full data reconstruction is shown again, for reference at bottom left, and the reconstruction using our limited angle algorithm is shown at bottom right. Notice that the edges are recovered in their proper places, although their magnitudes are somewhat lower. The only irritating artifact is the edge at the top right, which is somewhat enlarged. 16 Figure 7: Above we illutrate the reconstruction of the Shepp-Logan phantom from 145 degrees of noisy data. The full data reconstruction from this noisy data is shown at top left. The reconstruction using only the limited data, is shown at top right. The full data reconstruction is shown again, for reference at bottom left, and the reconstruction using our limited angle algorithm is shown at bottom right. Notice that the noise level is somewhat elevated, but that most of the structure is recovered. The noise is white gaussian, and the signal to noise ratio is 20:1. 17 Figure 8: Above we illutrate the reconstruction of the Shepp-Logan phantom from 90 degrees of data. The full data reconstruction from this noisy data is shown at top left. The reconstruction using only the limited data, is shown at top right. The full data reconstruction is shown again, for reference at bottom left, and the reconstruction using our limited angle algorithm is shown at bottom right. Although the reconstruction is not perfect, this is not surprising given that only half of the angular data was utilized. 18
© Copyright 2026 Paperzz