Open Access
3 February 2015 Lensless coded-aperture imaging with separable Doubly-Toeplitz masks
Michael J. DeWeert, Brian P. Farm
Author Affiliations +
Abstract
In certain imaging applications, conventional lens technology is constrained by the lack of materials which can effectively focus the radiation within a reasonable weight and volume. One solution is to use coded apertures—opaque plates perforated with multiple pinhole-like openings. If the openings are arranged in an appropriate pattern, then the images can be decoded and a clear image computed. Recently, computational imaging and the search for a means of producing programmable software-defined optics have revived interest in coded apertures. The former state-of-the-art masks, modified uniformly redundant arrays (MURAs), are effective for compact objects against uniform backgrounds, but have substantial drawbacks for extended scenes: (1) MURAs present an inherently ill-posed inversion problem that is unmanageable for large images, and (2) they are susceptible to diffraction: a diffracted MURA is no longer a MURA. We present a new class of coded apertures, separable Doubly-Toeplitz masks, which are efficiently decodable even for very large images—orders of magnitude faster than MURAs, and which remain decodable when diffracted. We implemented the masks using programmable spatial-light-modulators. Imaging experiments confirmed the effectiveness of separable Doubly-Toeplitz masks—images collected in natural light of extended outdoor scenes are rendered clearly.

1.

Introduction

Coded-aperture imaging was introduced by Dicke1 and Ables2 in the 1960s, and the field developed rapidly for astronomical x-ray and gamma-ray imaging. Coded-aperture imagers extend the pinhole camera concept, which requires no lenses, has unlimited depth of focus, and can image radiation of any wavelength. By placing a large number of pinholes in a common aperture plane, the light-gathering capability is greatly increased. Improved sensitivity comes at the price of having the focal plane record a multiplex of overlapping images, requiring algorithmic reconstruction to render a clear image. Thus, coded-aperture imaging is a subset of the general field of computational imaging.

Some limitations of a random pinhole distribution, primarily numerical instability due to the existence of zeros in the inverse-mask Fourier-domain transfer function, were mitigated with the development of coded apertures with uniformly redundant arrays (URAs)3 and modified URAs (MURAs)4 of pinholes. By placing the pinholes in prime-number-based patterns determined by sampling theory, it was thought that the near-delta-function system point-spread functions (PSFs) could be achieved (after deconvolving the distribution with appropriate filter functions), greatly improving the image reconstruction.

Use of coded-aperture imaging for general terrestrial applications poses two challenges. First, strictly speaking, the information-theoretic advantages of URAs and MURAs apply only for radiation wavelengths short enough that diffraction is negligible, since a diffracted MURA is no longer a MURA. Thus, coded-aperture imaging was applied mainly for x-ray and gamma-ray astronomies. Second, MURAs have limited effectiveness for extended scenes in general, regardless of radiation wavelength, due to the fundamentally ill-posed nature of reconstructing an extended scene from a finite-sized coded-aperture image. Thus, exploiting the potential of coded-aperture imaging requires addressing both diffraction and ill-posed reconstruction with very fast and efficient decoding methods. The work presented in this paper shows how to accomplish that goal.

1.1.

Objective: Lensless Imaging of Extended Scenes with Natural Illumination

In the astronomical applications for which coded aperture imaging was originally developed, the scene typically contains a discrete bright object in a largely dark field. A key assumption behind MURA encoding is that “all [nonzero] object points contribute a complete cyclic version of the basic [aperture] pattern,”3 and the imaging geometry is designed to ensure that the complete aperture “shadows” are collected on the focal-plane array. However, terrestrial applications involve nonzero scene elements extending to and past the edge of the effective field-of-view: these scene elements will cast only partial aperture shadows on the focal plane. This effectively causes more object points to contribute to the image than there are image pixels, creating an inherently ill-posed problem for reconstructing the “true” scene.

To address the ill-posed imaging problem, we developed a new class of coded aperture masks which provide compelling advantages over MURAs. The new class of masks, which we call separable Doubly-Toeplitz apertures, provides images that are orders of magnitude faster to decode and are less affected by the ill-posed nature of the imaging of extended scenes—their transfer functions have far fewer zero eigenvalues than do those of MURA masks. Unlike MURAs, separable Doubly-Toeplitz apertures retain their properties when diffracted, making them much more robust for broadband imaging than MURAs.

This paper addresses the optics and mathematics of the new class of coded apertures.

The method has also been validated with real-world imaging of extended scenes in natural light. Such validations are absolutely necessary—much of the published work based on simulations alone presents overly optimistic predictions of coded-aperture performance. Thus, we provide a demonstration of coded-aperture imaging in visible light acquired with off-the-shelf hardware. This work also uncovered several optical phenomena that are minor nuisances for lensed imaging, but causes significant performance degradation for coded apertures, and would not have been uncovered by modeling and simulation alone. A future publication will provide in-depth exposition of the practical optical challenges for developing software-defined optics.

1.2.

Former State-of-the-Art: Fourier Decoding and Modified Uniformly Redundant Array Masks

The coded-aperture mask is a flat plane parallel to an imaging array. In this configuration, the image I is just a correlation function between the object O and the mask M:

Eq. (1)

I=MO,
or in terms of matrix elements

Eq. (2)

I(i,j)klO(k,l)M(i+k,j+l).

The goal of decoding is to use knowledge of M to recover the object scene O from the encoded image I.

The prior state-of-the-art in coded-aperture imaging was to select binary codes which could be inverted via forward correlation with slightly modified versions of themselves. Specifically, mask codes M were selected for which the masks are simply a collection of binary (0% or 100% transparent) elements and for which the inverse matrix G, such that

Eq. (3)

OEstimated=GI=GMOO,
was easily derivable from M itself. For example, for URA3 masks,

Eq. (4)

Gi,j={+1,Mi,j=11,Mi,j=0.

URAs and MURAs satisfy Eq. (3) under periodic boundary conditions. In practice, diffraction is not always negligible. An example of an ideal mask is shown in Fig. 1, along with a diffracted mask image. In this case, the MURA was lithographically deposited as a pattern on a glass slide, so that it was truly binary, either passing or blocking light. However, the collected mask image shows effects of scattering and diffraction, compromising the self-inverse property. In addition, as is explained below, even ideal MURAs have limited utility for extended scenes.

Fig. 1

Example of a MURA: (a) ideal binary mask, and (b) image showing effects of diffraction. The collected mask image is no longer binary and exhibits numerous diffraction features.

OE_54_2_023102_f001.png

Because diffracted MURAs do not obey Eq. (3), inversion of the transfer function is required. The most straightforward approach for inversion of Eq. (1) is by Fourier transform. Our initial investigations using spatial-light-modulator (SLM)-defined masks showed noisy low-contrast recovered images. We successively eliminated the causes such as diffraction, polarization, and nonuniform SLM-element throughput. Finally, we traced the performance limitations to the ill-posed mathematics inherent in imaging extended scenes. The larger the mask, the more ill-posed the inversion, thus requiring regularization.

Initially, we implemented regularized decoding by introducing a real, positive function c2(k) into the Fourier-transform-based decoding and computing an estimated image from

Eq. (5)

ORecovered(k)I(k)M(k)/[|M(k)|2+c2(k)].

This is a special case of the Tikhonov5 regularization method used in image deblurring6c2(k) plays the dual roles of regularizing the inversion and controlling the effects of noise. With this regularization, recognizable decodings of laboratory images were achieved with both lithographically produced and programmable-SLM-based masks.

2.

New Approach to Ill-Posed Coded-Aperture Decoding: Theory

Even with regularized Fourier inversion, decoded images of extended scenes were usually unrecognizable. We traced this to the Fourier method itself, which has the limitation of implicitly assuming cyclic or periodic behavior of the object O. While the cyclic assumption is acceptable for compact objects against dark backgrounds, it is inappropriate for extended scenes viewed through large aperture masks.

To relax the assumption of cyclic boundaries, we recast the decoding problem in the most general terms. The correlation of a filter M with an object scene O to yield an image I in Eq. (1) can also be expressed as a simple matrix multiplication:

Eq. (6)

I1=M·O1,so thatI1i=kMi,kO1k,
where I1 and O1 are the one-dimensional (1-D) vectors created by concatenating the rows of I and O:

Eq. (7)

I1CI×i+j=Ii,j,andO1CO×i+j=Oi,j,
where CI and CO are the numbers of columns in I and O, respectively. The matrix M has rows that are shifted copies of the first row, which is constructed from the filter M by embedding each row of M into a segment of length CO, padded with zeros for elements CF+1 through CO, then concatenating the segments into a row of length CO×RO, where RO is the number of rows in O. That is,

Eq. (8)

M1,CO×i+j={Mi,jfor  jthe firstCFcolumns0for  jlater columns.

The size of the matrix M grows as the sizes of the image and the filter mask grow. If the image has dimensions rI×cI and the filter mask has dimensions rF×cF, then the sampled object is of size rO×cO, with rO=(rI+rF1) and cO=(cI+cF1). This yields for M dimensions of (rI×cI) rows and (rO×cO) columns.

The relationship between image size, filter size, and the extent of the object area sampled illustrates the inherently ill-posed nature of any imaging system. Since the PSF of a real-optical system is never a delta-function, there are always more object points contributing to an image than there are points in the image itself. For conventional imaging systems, where the PSF may be very compact, it is possible to clip or “throw-away” points near a reconstructed image’s edges and to retain only those which represent an approximately well-posed inversion. If significant diffraction is present, then the clipping strategy is inadequate—hence, the emphasis is on the diffraction limit in conventional optics.

For coded-aperture imaging, the filter area (and the spread function) is directly proportional to the light-gathering power of the system. Because of this proportionality of spread to light-gathering, it is not possible to use a clipping strategy to select well-posed portions of the recovered image, except in the trivial case of a single-pinhole aperture. Thus coded-aperture imaging must address inherently ill-posed encoding and decoding.

While the forward correlation can be done exactly using Eq. (6), the inversion (estimating O from the measurements I) presents a seriously ill-posed7 and computationally challenging task. The general case is commonly solved via the regularized normal equations:

Eq. (9)

MTI1=(MTM+α21rO×cO)O1,
where 1rO×cO is the identity matrix with each dimension measuring

Eq. (10)

rO×cO=(rI+rF1)×(cI+cF1).

The operations count for Eq. (9) is on the order of8

Eq. (11)

NOperations,Generalized(rOcO)3.

For 1 megapixel images, this yields 1018 operations, requiring on the order of 106s to 12 days on a teraflops processor. Thus, computations on large images using generalized masks are usually impractical. Further, the matrix M is so large that the accumulation of round-off errors will likely swamp the inversion even if it could be accomplished in a manageable time. The only exception would be if the encoding mask is a single-pixel pinhole aperture—a well-posed but trivial case.

2.1.

Separable Doubly-Toeplitz Masks

We have developed a method to specify finite-sized masks that greatly reduce the degree to which the imaging is ill-posed and render the inversions of Eq. (9) tractable. If the mask M can be expressed as an outer product of two 1-D vectors A and B, with lengths NA and NB, respectively, then M is an NA×NB-dimensional matrix of the form

Eq. (12)

M(i,j)=A(i)B(j),
where A(i) is the i’th element of vector A, and B(j) is the j’th element of vector B. Equation (9) can now be written simply as a product of much smaller two-dimensional (2-D) matrices:

Eq. (13)

I=MAOMBT.

The matrices MA and MB are Toeplitz, with the elements shifted by 1 pixel on each successive row:

Eq. (14)

MA=[A1A2ANA00000A1A2ANA000].

The vector MB is similarly constructed from the elements of the vector B.

The difficulty of inversion using Eq. (13) is vastly less than for solving Eq. (9). While this is still an ill-posed problem, the sizes of matrices MA and MB are much smaller than M in Eq. (9). With the matrix dimensions summarized in Table 1, the inversions require an operations count on the order of

Eq. (15)

NOperations,Doubly-Toeplitz[(rIrO)3+(cIcO)3].

Table 1

Relationship of dimensions of the collected image, the encoding filter mask, and the object area contributing to the image.

MatrixRowsColumns
Coded-aperture mask filter, MrFcF
Image on focal plane, IrIcI
Object area, OrO=(rI+rF1)cO=(cI+cF1)
Left Toeplitz mask, MArIrO
Right Toeplitz mask, MBcIcO

For image sizes on the order of 1 megapixel, this yields on the order of 2×109 operations, requiring just milliseconds on a teraflops processor. Thus, unlike arbitrary general masks or diffracted MURAs, separable Doubly-Toeplitz masks can be inverted in a usefully short time.

To our knowledge, this Doubly-Toeplitz strategy for specifying coded apertures has not been previously applied to the spatial imaging problem, though there are hints of 1-D Toeplitz masks in the temporal motion-deblurring work of Agrawal and Raskar.9,10

3.

Diffraction Effects on Doubly-Toeplitz Masks

Having identified an approach to reduce the impact of ill-posed inversions, the investigations next turned to mitigating the effects of diffraction. Simulations verified that the Doubly-Toeplitz property is preserved in a diffracted mask. Simulations also provide a means of synthesizing decoding matrices representing diffracted mask images for use when good laboratory-collected mask images are not available.

Diffraction simulations are particularly straightforward for a collection of identical rectangular apertures placed at points on a regular-rectangular-grid SLM. Let IO and IP denote the intensity fields at the object and image planes, respectively, and let EO denote the complex optical field at the object plane. When the illumination is incoherent, the diffracted intensity in a pixel at location (i,j) on the image plane is

Eq. (16)

IP(i,j)k,l|QX(i+k)|2IO(k,l)|QYT(l+j)|2,
where

Eq. (17)

QX(kx)=jXA(jX)exp{i2πλjXpX[kxΔxPfPjXpX2(1R+1fP)]}×sinc{πλaX[kxΔxPfPjXpX(1R+1fP)]}
and

Eq. (18)

QY(ky)=jYB(jY)exp{i2πλjYpY[kyΔyPfPjYpY2(1R+1fP)]}×sinc{πλaY[kyΔyPfPjYpY(1R+1fP)]}.

The parameters in Eqs. (16)–(18) are

(aX,aY) = aperture opening sizes,

(pX,pY) = aperture pitch, >(aX,aY),

(ΔxP,ΔyP) = image-plane pixel pitch,

A, B = vectors of weights of mask elements in the x and y directions,

R = range from aperture mask to object plane,

fP = range from aperture mask to image plane, and

λ = wavelength of light.

Diffraction is a significant contributor to extended-scene coded-aperture imaging in visible light. Our experimental investigations were conducted with a bandpass filter centered on 550 nm, and a coded aperture implemented by varying the throughput of elements of a transmissive SLM. The SLM had elements with dimensions of approximately 24μm×31μm on a 36-μm pitch. The total area of active pixels measured 1024×768pixels, corresponding to a maximum aperture of approximately 27mm×36mm. The distance from the aperture to the focal plane was typically 65 mm. With these parameter values, the Fresnel number for the coded aperture optics ranged from under 0.02 (with a single SLM element used as a pinhole) to over 40,000. Because the objective of coded-aperture imaging is to capture the advantages of pinhole imaging while maximizing light-gathering capability, many SLM elements were set to the “on” state, so that our experiments were conducted predominantly at Fresnel numbers between 150 and 40,000. We simulated the Fresnel diffraction with a custom-written numerical code developed for coded-aperture imaging and validated by comparison with laboratory results using both programmable and fixed apertures. Results of one of these experiments are shown in Fig. 2. Within the experimental uncertainties caused by random variations of the transmissivity of the SLM elements and the finite resolution of the focal plane, the agreement is very good. In future papers, we plan to present the effects of diffraction, noise, and variable transmissivity on the quality of coded-aperture images via objective and subjective metrics such as peak signal-to-noise ratio and structural similarity.

Fig. 2

Images of a point source viewed through a single-large square aperture synthesized by setting a 100-pixel-square area of SLM pixels to the “on” state. Vertical cuts through the images illustrate the diffraction features. The agreement is very good, despite noise and the element-to-element variability of the SLM.

OE_54_2_023102_f002.png

Examination of Eq. (16) shows that, if the raw aperture mask is an outer product of two 1-D vectors A and B, then the effects of the diffracted mask can still be expressed as an operation of a pair of Toeplitz matrices. Thus, a separable Doubly-Toeplitz aperture mask will yield a separable Doubly-Toeplitz encoding even when there is significant diffraction. If the object is an incoherent point-like source, then its image will be decomposable into A and B vectors which can be used to synthesize decoding masks.

To verify the diffraction model, we conducted a series of experiments comparing simulations with image data, obtaining excellent agreement between model and data. Diffracted coded apertures were also modeled and compared with “collected masks”—images of a point-source viewed through the spatial-coding masks. We verified that the collected masks decompose into the outer products of diffracted 1-D functions, in accord with Eq. (16). The mask shown in Fig. 3 is the basic “workhorse” mask used in the investigations presented in this paper. Three scaled versions of this mask were used: “02” is 2.2-mm wide, “04” is 4.4-mm wide, and “08” is 8.8-mm wide. The “02” mask spans 62 elements of our SLM, with a minimum feature size of two SLM elements. The SLM elements had a nominal 35.5-μm pitch. The pitches were identical along the axes of the SLM, but the aperture openings were asymmetric—24-μm high and 31-μm wide, accounting for the asymmetry of the diffraction pattern. The 31-element base mask is adapted from the one published by Agrawal and Raskar10 for temporal encoding and decoding to compensate for motion blur. In a future publication, we will report on methods for efficient searches for masks optimized specifically for 2-D spatial coded-aperture imaging.

Fig. 3

(a) Binary mask commanded on the SLM, and (b) the image of a point source viewed through the mask, showing Fresnel diffraction features.

OE_54_2_023102_f003.png

Figure 3 also demonstrates that the diffraction increases the size of the effective encoding masks, increasing the ill-posedness of the inverse problem. This makes separable Doubly-Toeplitz masks even more valuable for cases in which fine imaging resolution is required.

4.

Solution: Method

In real-imaging systems, noise will always arise from effects such as photon counting or read noise, so that the actual inverse problem includes a noise term N. The matrix Iδ denotes the image containing noise effects which are not encoded by the aperture mask. In addition, there is the ill-posed nature of the problem embodied in the nonsquare matrices MA and MB. Both the noise and the ill-posedness require a regularization method. We begin by multiplying Eq. (13) by the transposes of the mask matrices, obtaining the normal equation:

Eq. (19)

MATIδMB=(MATMA)O(MBTMB)T.

The matrix products (MATMA) and (MBTMB) are square and symmetric, but are not invertible, due to degenerated zero eigenvalues. The degree of degeneracy increases with the size of the filter masks—there are (rOrI) zero eigenvalues for MA, and (cOcI) zero eigenvalues for MB. [Compare this with solving Eq. (9) for a MURA or generic mask, which is orders of magnitude more degenerate, having (rOcOrIcI) zero eigenvalues.]

To relieve the degeneracies, we introduce regularization parameters αA and αB, so that

Eq. (20)

MATIMB(MATMA+αA21RO)O(MBTMB+αB21CO)T+MATNMB,
where 1RO is an identity matrix with dimensions RO, and 1CO is an identity matrix with dimensions CO. Equation (20) now multiplies O by invertible matrices, so that a noisy estimate of O can be derived from the image:

Eq. (21)

OEst=(MATMA+αA21RO)1MATIMB(MBTMB+αB21CO)1.

5.

Methods for Improving the Solution

5.1.

Normalization of Reconstructed Images

The reconstructed images initially show patterns at their edges within a distance of one mask width (or height) of the border. We noted similar artifacts in the encoded images of uniform (all white) scenes, which led us to a method by which the artifacts could be normalized out of the reconstructed images. The normalization is accomplished by encoding and decoding a uniform all-white object W to create a matrix WEst that includes the spurious brightness differences accrued as the coded-aperture mask straddles the edge of the image area:

Eq. (22)

WEst=[MATMA+αA21RO×RO]1(MATMA)W(MBTMB)[MBTMB+αB21CO×CO]1.

OEst is divided by WEst, element-by-element, to obtain a normalized object estimate ONorm:

Eq. (23)

ONorm(i,j)=OEst(i,j)/WEst(i,j).

The normalization can also be expressed as multiplication by two normalization matrices DA and DB:

Eq. (24)

ONorm=DAOEstDBT.

This is because the all-white matrix W can be expressed as an outer product of two 1-D vectors of constant value, which in turn allows the normalization estimate to be decomposed into an outer product of two 1-D matrices:

Eq. (25)

WEst(i,j)=WA(i)WB(j).

The normalization matrices DA and DB each have identical columns, given by

Eq. (26)

DA(i,j)=1/WA(i)=1/{k=1RO[(MATMA+αA21RO×RO)1(MATMA)](i,k)}DB(i,j)=1/WB(i)=1/{k=1CO[(MBTMB+αB21CO×CO)1(MBTMB)](i,k)}.

The decoded and normalized object estimates are then

Eq. (27)

ONorm=DA(MATMA+αA2)1MATIδMB[(MBTMB+αB2)T]1DBT.

5.2.

Improvement via Landweber Iteration

The results can be further improved by iterating11 from the solution in Eq. (27). We settled on the method of Landweber,12 which iterates using a regularization constant εL:

Eq. (28)

O(n+1)=O(n)+εLMATPA[IδMAO(n)MBT]PBTMB,
where PA and PB are the positive-definite symmetric matrices specified to control the convergence. While classic Landweber implementation uses PA=PB=1 (the identity matrix), our implementation uses

Eq. (29)

PA=DASAandPB=DBSB,
where SA and SB are the inverses:

Eq. (30)

SA=(MATMA+αA2)1,andSB=(MBTMB+αB2)1.

The starting estimate O(0) is estimated from Eq. (27), then iterated using

Eq. (31)

O(n+1)=O(n)+εL2DASAMAT[IδMAO(n)MBT]MBSBDB+an(O(n)O(0)),
with

Eq. (32)

an=1/(4+n).

Equation (31) is actually a modified Landweber method13 that is guaranteed to converge as long as 0<εL2<2. However, the convergent solution is not guaranteed to be correct—the results may be noise dominated unless the iterations stop at a suitable point. We apply an a posteriori criterion based on the residuals after n iterations:

Eq. (33)

R(n)=MAT[IδMAO(n)MBT]MB,
with norm

Eq. (34)

r(n)2=i,j[R(n)(i,j)]2.

The algorithm stops the iterations when the residuals from Eq. (34) begin to increase.

6.

Encoding and Decoding Real-Extended Imageries

The refined methods of coding and decoding made it possible to image extended outdoor scenes in natural light. We used these images to further improve and debug the processing.

One experimental setup is shown in Fig. 4. The camera was situated in a commercial high-rise building in Honolulu, Hawaii, approximately 45 m above the ground, looking east toward the city skyline. The SLM was a Holoeye HEO 0017 LC-SLM with a 1024×768-element active area. This was paired with a Kodak RMV 16000 digital camera. The RMV 16000 was chosen because its focal plane area of 36mm×24mm is close in size to the 33mm×25mm dimensions of the Holoeye SLM and because the RMV pixel pitch of 7.4μm allows oversampling of the nominal 35.5-μm pitch of the SLM elements, reducing sampling as a source of error. The SLM was placed 6.5 cm in front of the camera’s focal plane.

Fig. 4

Equipment for imaging outdoor scenes. This view looks from downtown Honolulu eastward toward Diamondhead mountain.

OE_54_2_023102_f004.png

In addition to the programmable SLM, the camera had a green-bandpass filter to reject optical wavelengths that are not adequately blocked by the “off” pixels of the SLM. We also developed methods to deal with “bad” or “dead” pixels, inhomogeneous focal plane response, fixed-pattern noise, and with stray light reflecting inside the camera. Most of the effects are minor nuisances for conventional lens-assisted imaging, but cause significant degradation of coded-aperture decoding quality. This is because the decoding algorithms attempt to invert the image, artifacts and all, to construct an object scene. Careful reduction of all nonencoded-light is required to achieve satisfactory results.

6.1.

Encoding and Decoding Real-Extended Imagery: Results

The details of the various artifacts and the associated mitigation algorithms will be presented in a future publication. The basic steps to precondition the images then to decode them are outlined in Fig. 5. The denoising step can be either single-frame denoising or denoising by adding multiple raw images before decoding. Stray-light interference limited the effectiveness of single-frame denoising. This was mitigated via a 20-frame average, with the raw images averaged together before application of flat, dark, and defect corrections.

Fig. 5

Processing flow for coded-aperture image acquisition and decoding. Raw coded-aperture imagery is unrecognizable—processing yields a clear image. Data conditioning reduces numerous sources of unencoded noise and interference, so that the decoding algorithms can compute a recognizable image.

OE_54_2_023102_f005.png

6.2.

Multimask Image Stacking

Our investigations showed that the regularization artifacts differ from mask to mask. Thus, reconstruction could be improved by using multiple images collected with a sequence of different masks. We tested this approach using four 90-deg rotations of the same basic mask. The four images were decoded separately and stacked to provide an improved image. Results from simply averaging the four resulting decoded images are compared with single-mask imagery in Fig. 6 and also compared with an image acquired with a conventional lensed camera.

Fig. 6

Final results (b) of processing real imagery with multiple masks (four 90-deg rotations of the “02” mask), versus using a single mask (a), showing reduction of decoding “ghost” artifacts. (c) The reference photo of the scene acquired with a conventional camera. (The clouds are different because the images were acquired at different times.)

OE_54_2_023102_f006.png

7.

Conclusions

Our goal was to improve the performance of completely lensless imagers, especially for imaging extended scenes in natural light. We have demonstrated a novel approach, separable Doubly-Toeplitz masks, for specifying coded-aperture masks that allow efficient decoding and have shown the feasibility of using programmable SLMs to implement the new class of masks. This work presents a fundamental change in coded-aperture imaging and lays the foundation for future developments in which we will quantify the performance of the separable Doubly-Toeplitz masks for a greater variety of scenes with both programmable and fixed masks, exploring the limitations of the undistorted field-of-view achievable, as well as extending the applicability to bands beyond the visible spectrum.

In particular, it has been suggested that the diffraction could limit the performance of coded apertures for infrared imaging.14 The work presented here shows that the diffraction can be readily included in the decoding algorithms for separable Doubly-Toeplitz masks. More serious limiters, especially for the long-wave infrared, are likely to be the influence of nonencoded information from thermal self-emission of the camera body and drifts in the nonuniformity corrections for uncooled thermal-camera focal planes.

Future work will also explore how the separable-Doubly-Toeplitz masks can be combined with refractive elements, extending the super-resolution capabilities studied under the DARPA LACOSTE program.15

The project has also identified other implementation-specific difficulties that need to be addressed in using SLMs for lensless imaging. Since it is still early in the development of the technology, rapid progress can be made. Once the device-specific sources of noise and error are mitigated, the next biggest improvement will come from developing a computationally efficient means for identifying mask candidates. Rather than generate random binary codes, then exhaustively test each one, a more-efficient approach is likely to be a genetic algorithm which evolves mask codes from the best currently known starting points.

Since Doubly-Toeplitz masks are continuous, having no isolated open sections, self-supporting nonprogrammable masks can be built in to the fuselages and carapaces of small unmanned vehicles with three-dimensional printing methods, reducing the time and cost of fielding new systems. If fabricated from material that blocks x-ray and gamma-ray radiations, then this would in turn enable extremely compact noncontact imaging in these wavelengths for medicine, defense, and homeland security.

Acknowledgments

We are grateful for support for this work provided by the Office of Naval Research under contract N00014-10-C-0193.

References

1. 

R. H. Dicke, “Scatter-hole cameras for x-rays and gamma rays,” J. Astrophys., 153 L101 –L106 (1968). http://dx.doi.org/10.1086/180230 JASRD7 0250-6335 Google Scholar

2. 

J. G. Ables, “Fourier transform photography: a new method for x-ray astronomy,” Proc. Astron. Soc. Aust., 1 (4), 172 –173 (1968). AAUPBC 0066-9997 Google Scholar

3. 

E. E. FenimoreT. M. Canon, “Coded aperture imaging with uniformly redundant arrays,” Appl. Opt., 17 (3), 337 –347 (1978). http://dx.doi.org/10.1364/AO.17.000337 APOPAI 0003-6935 Google Scholar

4. 

S. R. GottesmanE. E. Fenimore, “New family of binary arrays for coded aperture imaging,” Appl. Opt., 28 (20), 4344 –4352 (1989). http://dx.doi.org/10.1364/AO.28.004344 APOPAI 0003-6935 Google Scholar

5. 

A. N. Tikhonovet al., Numerical Methods for the Solution of Ill-Posed Problems, Kluwer Academic Publishers, Moscow, Russia (1995). Google Scholar

6. 

S. Kimitei, Algorithms for Toeplitz Matrices and Their Applications to Image Reconstruction, Georgia State University, Atlanta, Georgia (2005). Google Scholar

7. 

J. Hadamard, “Sur les problèmes aux dérivées partielles et leur signification physique,” Princeton Univ. Bull., 13 (4), 49 –52 (1902). Google Scholar

8. 

W. H. Presset al., “Highly-optimized algorithms may require just N^(log27) operation instead of N^3 operations,” Numerical Recipes in C—The Art of Scientific Computing, 102 –104 2nd ed.Cambridge University Press, New York (1999). Google Scholar

9. 

R. RaskarA. AgrawalJ. Tumblin, “Coded exposure photography: motion deblurring using fluttered shutter,” ACM Trans. Graphics (TOG) Proc. ACM SIGGRAPH, 24 (3), 795 –804 (2006). http://dx.doi.org/10.1145/1179352.1141957 ATGRDF 0730-0301 Google Scholar

10. 

A. AgrawalR. Raskar, “Resolving objects at higher resolution from a single motion-blurred image,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition ‘CVPR’, 1 –8 (2007). Google Scholar

11. 

W. H. Presset al., Numerical Recipes in C—The Art of Scientific Computing, 55 –58 2nd ed.Cambridge University Press, New York (1999). Google Scholar

12. 

T. ElfvingT. Nikazad, “Stopping rules for Landweber-type iteration,” Inverse Probl., 23 1417 –1432 (2007). http://dx.doi.org/10.1088/0266-5611/23/4/004 INPEEY 0266-5611 Google Scholar

13. 

H. Egger, “On the convergence of modified Landweber iterations for nonlinear inverse problems,” (2010). Google Scholar

14. 

S. R. Gottesman, “Coded apertures: past, present, and future application and design,” Proc. SPIE, 6714 671405 (2007). http://dx.doi.org/10.1117/12.735494 PSISDG 0277-786X Google Scholar

15. 

A. Mahalanobiset al., “IR performance study of an adaptive coded aperture “diffractive imaging” system employing MEMS “eyelid shutter” technologies,” Proc. SPIE, 6714 67140D (2007). http://dx.doi.org/10.1117/12.745956 PSISDG 0277-786X Google Scholar

Biography

Michael J. DeWeert is a BAE Systems Engineering Fellow and the chief technology officer at BAE Systems in Honolulu. He has over 25 years of experience in remote sensing research and development for intelligence, surveillance, and reconnaissance. His current responsibilities include technical oversight spectral imaging, LIDAR, interferometric remote-sensing systems, and modeling and simulation for maritime and terrestrial sensing. Currently, he is the principal investigator for projects developing advanced imaging technologies for long-standoff mine countermeasures.

Brian P. Farm has over 25 years of experience in modeling, algorithm/software development, and spatial and image data analyses. He is involved in algorithm/software research and development for data exploitation, real-time image analysis, and sensor-model adjustment and exploitation. He has worked on 3-D imaging LIDARs, developed registration and processing algorithms for imaging and interferometric systems, and developed information processing for novel, nontraditional imaging sensors, and created simulation systems to allow advanced algorithm testing on proposed new sensor systems.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Michael J. DeWeert and Brian P. Farm "Lensless coded-aperture imaging with separable Doubly-Toeplitz masks," Optical Engineering 54(2), 023102 (3 February 2015). https://doi.org/10.1117/1.OE.54.2.023102
Published: 3 February 2015
Lens.org Logo
CITATIONS
Cited by 68 scholarly publications and 4 patents.
Advertisement
Advertisement
KEYWORDS
Diffraction

Spatial light modulators

Coded apertures

Photomasks

Image filtering

Matrices

Imaging systems

Back to Top