Team develops camera that uses sensors with just 1,000 pixels

Carnegie Mellon researchers develop camera that uses sensors with just 1,000 pixels
Credit: http://www.ece.cmu.edu/_media/images/lisense_dmd_sensor.png

Thanks to the `megapixel wars', we are used to cameras with 10s of megapixels. Sensors in our cell phone and SLRs are made of Silicon (Si), which is sensitive to the visible wavebands of light and hence, useful for consumer photography. The abundance of Silicon, coupled with advances in CMOS-based fabrication has helped drive down the cost of sensors while simultaneously providing increased capabilities in terms of sensors with higher and higher resolutions. 

In many wavebands that are outside Si's sensitivity, sensing can be very expensive. Two examples of these are in short-wave infrared (0.9-3 microns) and mid-wave infrared (3-8 microns); the cost of a megapixel sensor in both wavebands is typically in 10s of thousands of dollars. As a consequence, high-resolution short-wave and mid-wave are too expensive to be affordable for the average consumer.

Researchers at CMU and Columbia University have developed a new camera, called LiSens, that uses a sensor with just a thousand pixels, but produces images and videos at (nearly) a mega-pixel resolution. In other words, LiSens takes a low-resolution sensor and by the use of novel optic makes it capable of sensing scenes at a resolution that is higher than that of the sensor. This is achieved by focusing the scene onto a digital micro-mirror array (DMD) and, subsequently, focusing the DMD onto the low-resolution sensor. The DMD is an array of tiny mirrors that can direct light towards or away from the sensor.

Carnegie Mellon researchers develop camera that uses sensors with just 1,000 pixels
Credit: http://www.ece.cmu.edu/_media/images/lisense_dmd_sensor.png

In the case of LiSens, we use a linear array of sensing elements or a line-sensor such that each pixel on the sensor can sum together light on an entire column of the DMD. Given the programmable nature of the DMD and its ability to block light at desired pixels, each measurement obtained by the line-sensor is a coded sum along a line in the scene. Given multiple such measurements with different codes, we can recover the image focused on the DMD at its full-resolution.

A measurement obtained by the line-sensor is quite unlike a traditional image. However, we can design algorithms that, armed with the knowledge of the scene-to-sensor mapping, can invert these measurements obtained by the line-sensor and compute an image of the scene. With this, we can now sense scenes at high-resolution (in our prototype, 1024x768 pixels) in spite of having a low-resolution sensor (in our prototype, 1024 pixels).

LiSens builds on the so-called single pixel camera (SPC), which uses a single photodetector to sense the scene. LiSens can be interpreted as a multi-pixel extension of the SPC. LiSens delivers measurement rates that are nearly 100-1000 times that of the SPC; this allows LiSens to sense scenes at significantly higher spatial and temporal resolutions. The price that we pay is a moderate increase in the cost of the sensor.

Aswin Sankaranarayanan's work was published in The 2015 International Conference on Computational Photography held in Houston, TX.


Explore further

Banknote check with ultra-fast line scan sensor

Citation: Team develops camera that uses sensors with just 1,000 pixels (2015, June 2) retrieved 21 October 2019 from https://phys.org/news/2015-06-team-camera-sensors-pixels.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
2908 shares

Feedback to editors

User comments

Jun 04, 2015
It took a whole lot of words to say that they're simply using the micromirror array to let the line sensor see only parts of the scene at a time, thereby making it work like a scanning camera back without physically moving the sensor.

There's all sorts of statistical tricks you ca do with patterns when you do that, because you don't have to scan the image in sequence. The single pixel camera works that way by measuring ever finer and finer random patterns and summing the measurements up so you start from a single huge pixel and then as the noise adds up it starts to reveal details.

Ultimately you end up with a pattern that blinks individual pixels one at a time to the sensor to get the last absolutely accurate reading, but you can stop at any point before that if you feel the result is good enough, thus saving a lot of time.


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more