Analog chip sorts colors

April 1, 2002
Researchers at Iguana Robotics (Mahomet, IL) and Johns Hopkins University (Baltimore, MD) have demonstrated a single-chip analog image-processing system that can acquire color images, perform color segmentation, transform image color space, create color profiles, and match acquired images with templates.

Researchers at Iguana Robotics (Mahomet, IL) and Johns Hopkins University (Baltimore, MD) have demonstrated a single-chip analog image-processing system that can acquire color images, perform color segmentation, transform image color space, create color profiles, and match acquired images with templates.1 Because such devices are very small, low in power consumption, and inexpensive to manufacture compared to their digital counterparts, their applications should be many, including microrobotics, personal computers, and toys that "see."

This work, presented to the IEEE International Solid-State Circuits Conference in San Francisco in February, is an extension of research done at California Institute of Technology (Cal Tech; Pasadena, CA) in the mid-1990s. The Cal Tech team showed that it was more appropriate to use hue-saturation- intensity (HSI) color space rather than the red-green-blue (RGB) coordinate system, and demonstrated an analog circuit that could translate between the two. Their approach showed how the HSI color space, which is most commonly used to understand human vision system, produces feature sets that are better for image segmentation: breaking the image into individual sections that can be considered a discrete object and then pattern-matched. Their analysis showed that, in some circumstances, HSI-based coordinate systems are less variant to some changing lighting conditions—such as shading, highlighting and transparency—than their RGB counterparts.

The 128 x 64 pixel low-resolution image is based on the look-up calculation of integrated mixed-signal circuitry. The pixel hue is calculated and sorted into discrete values or bins that are 10° apart. A Rubik's cube image is presented to the chip. First, the chip acquires an RGB image of it (left). Through the RGB-HSI transformation procedure, the image is segmented into pixels of different hue. In one example, areas are identified as being a particular yellow (center). All highlighted pixels are segmented to the same bin. As the pixels are redefined, a map of the image associated with hue can be produced (right).

The Iguana Robotics chip consists of a 128 x 64 (horizontal by vertical) array of imaging pixels, which sample RGB signals sequentially (using a color wheel in this prototype) and holding them for further use. Integrated with these detectors are mixed-signal electronics to perform the various processing steps. Processing is performed in blocks, the size and scanning patterns for which can be programmed in for a given application, the first step being the RGB-HSI transformation. This operation is performed by summing the RGB components to get intensity, subtracting the minimum of the normalized RGB values from this to get the saturation, and using an analog version of a lookup table to implement the equation for hue. This latter circuit also maps the RGB data into 36 hue intervals, each spaced 10° apart.

Once the color space has been converted, a profile of the image is created based on its HSI histogram. The color space is divided into several discrete areas and the number of pixels within each chunk counted (see figure). In the learning phase, image blocks of particular interest are identified and the camera takes this information, develops the appropriate profile, and stores it in an array of static random-access memory cells. In the pattern-matching phase, the block-pattern profiles from the image currently being scanned are compared with the templates for these stored regions of interest and the error codes generated used to find the best match. The researchers say that, using the RGB filter technology commonplace with conventional color cameras, the footprint of their imaging and recognition will be tiny: their current device has an area of less than 0.25 cm2.

REFERENCE
1. Ralph Etienne-Cummings et al, IEEE Int. Solid-State Circ. Conf., San Francisco, CA, 4 (February 2002).

About the Author

Sunny Bains | Contributing Editor

Sunny Bains is a contributing editor for Laser Focus World and a technical journalist based in London, England.

Sponsored Recommendations

What are Notch Filters?

Feb. 27, 2025
Notch filters are ideal for applications that require nearly complete rejection of a laser line while passing as much non-laser light as possible.

Using Optical Filters to Optimize Illumination in Fluorescence and Raman Systems

Feb. 27, 2025
Discover how Semrock products can help you get the most out of your fluorescence and Raman excitation designs, regardless of light source.

Melles Griot Optical Systems and Semrock Optical Filters for Spatial Biology

Feb. 26, 2025
Discover why a robust, high-throughput fluorescence imaging system with Semrock optical filters is key for Spatial Biology.

Understanding Practical Uses and Optimization Techniques for Fluorescence Optical Filters

Feb. 26, 2025
Learn about optical fluorescence and which optical filters to include in your instrument set up. See more with Semrock filter sets.

Voice your opinion!

To join the conversation, and become an exclusive member of Laser Focus World, create an account today!