Digital Imaging: Plenoptic camera refocuses digitally

Jan. 1, 2006
A digital handheld camera developed by researchers at Stanford University looks to all the world like an ordinary consumer camera, but has capabilities that are anything but ordinary.
A handheld plenoptic camera captures an image that, when digitally processed, can be refocused from near (top) to far (center), or even a range encompassing both (bottom). These images were upsampled by a linear factor of two to smooth them (the pixels have low noise so they tolerate upsampling well).
A handheld plenoptic camera captures an image that, when digitally processed, can be refocused from near (top) to far (center), or even a range encompassing both (bottom). These images were upsampled by a linear factor of two to smooth them (the pixels have low noise so they tolerate upsampling well).

A digital handheld camera developed by researchers at Stanford University looks to all the world like an ordinary consumer camera, but has capabilities that are anything but ordinary.1 For example, the image produced by the camera (or, more precisely, by its associated digital processor) can be reprocessed at any time to move the plane of best focus forward and back, as if refocusing the lens. More startling, the image can be processed in a slightly different way to create sharp focus everywhere-in effect, extending the depth of field way beyond what one would expect for the aperture stop (f-stop) of the lens.

The camera is a modification of the so-called “plenoptic” camera invented by John Wang and Edward Adelson at the Massachusetts Institute of technology (MIT; Cambridge, MA) in 1992.2 In a plenoptic camera, an ordinary camera lens creates an image in the usual way, but rather than placing a sensor array exactly at the image plane, a microlens array is placed there instead, with each microlens becoming an image pixel. The actual image sensor-which must have a much higher pixel count than the microlens array-is placed a short distance behind the microlens array; each microlens in the array images the exit pupil of the camera lens onto a unique sensor area consisting of, perhaps, a few hundred sensor pixels.

Although the number of lenses in the microlens array defines the number of pixels in the final image, the small exit-pupil image each microlens creates on the sensor array carries a multitude of data points on that single image pixel. In fact, each of the hundreds of sensor pixels (the data points) devoted to a given image pixel carries information on a different perspective for that image pixel, and thus for the pixel’s equivalent at the object plane (which is the portion of the exterior scene on which the camera lens is focused).

Wang and Adelson intended the plenoptic camera to be a device that would capture a single image, but allow the viewer to select a subset of sensor pixels (through digital processing) that would allow the viewer to alter the perspective of the image, and thus peek around 3-D objects.

The Stanford version

The plenoptic camera created by the Stanford group differs in two ways from the original version. First, its optical path is simplified, allowing it to be integrated into an ordinary (though high-pixel-count) digital still camera. Second, additional processing approaches have been developed to allow digital refocusing and digital depth-of-focus enhancement (see figure). As with the original version, the Stanford camera allows shifts in perspective (an ability that works best for macro photography, in which the camera lens is close to a small object and relatively large in size in relation to it).

The Stanford camera includes a 4000 × 4000-pixel sensor array with a 9-µm pitch, a 296 × 296-microlens array with lenslets of 500-µm focal length and 125‑µm diameter, and either a 140-mm f/2.8 or an 80-mm f/2.0 camera lens, both with adjustable stops. The software produces a 4-D light field (two lateral dimensions arising from the image plane and two from the pupil plane). Any vignetting is normalized by dividing each image-pixel output by the fraction of rays intercepted by the sensor light field for that image pixel.

With a camera lens set to an f/4 stop, for example, the technique leads to focal depths akin to those of higher f-stops. “With N×N sensor pixels under each microlens, you can theoretically refocus from an f/4 (say) to an f/(4N),” says Ren Ng, one of the researchers. “In our case, we find empirically that you can focus from f/4 to f/22. The f/4 version collects all the light coming through the f/4 aperture, which is about 36 times more than an f/22 aperture.”

Data sets captured by the plenoptic camera could be used for entertainment; a viewer could sit at a computer and change image perspectives, focal depths, and focal positions at will. These data may also have more-serious uses-for example, allowing the capture of a transient phenomenon at large focal depth and from simultaneous differing perspectives. Ng notes that the Stanford group has printed a computer-generated hologram from one of the data sets.

REFERENCES

1. R. Ng et al., Tech. Rep. CSTR 2005-02, Stanford Computer Science; graphics.stanford.edu/papers/lfcamera.

2. T. Adelson and J.Y.A. Wang, IEEE Trans. Pattern Analysis and Machine Intell. 14(2) (February 1992).

About the Author

John Wallace | Senior Technical Editor (1998-2022)

John Wallace was with Laser Focus World for nearly 25 years, retiring in late June 2022. He obtained a bachelor's degree in mechanical engineering and physics at Rutgers University and a master's in optical engineering at the University of Rochester. Before becoming an editor, John worked as an engineer at RCA, Exxon, Eastman Kodak, and GCA Corporation.

Sponsored Recommendations

Request a quote: Micro 3D Printed Part or microArch micro-precision 3D printers

April 11, 2024
See the results for yourself! We'll print a benchmark part so that you can assess our quality. Just send us your file and we'll get to work.

Request a free Micro 3D Printed sample part

April 11, 2024
The best way to understand the part quality we can achieve is by seeing it first-hand. Request a free 3D printed high-precision sample part.

How to Tune Servo Systems: The Basics

April 10, 2024
Learn how to tune a servo system using frequency-based tools to meet system specifications by watching our webinar!

Precision Motion Control for Sample Manipulation in Ultra-High Resolution Tomography

April 10, 2024
Learn the critical items that designers and engineers must consider when attempting to achieve reliable ultra-high resolution tomography results here!

Voice your opinion!

To join the conversation, and become an exclusive member of Laser Focus World, create an account today!