Improving image sensors for machine vision
Image sensors measure light intensity, but angle, spectrum, and other aspects of light must also be extracted to significantly advance machine vision.
In Applied Physics Letters, researchers at the University of Wisconsin-Madison, Washington University in St. Louis, and OmniVision Technologies highlight the latest nanostructured components integrated on image sensor chips that are most likely to make the biggest impact in multimodal imaging.
The developments could enable autonomous vehicles to see around corners instead of just a straight line, biomedical imaging to detect abnormalities at different tissue depths, and telescopes to see through interstellar dust.
"Image sensors will gradually undergo a transition to become the ideal artificial eyes of machines," co-author Yurui Qu, from the University of Wisconsin-Madison, said. "An evolution leveraging the remarkable achievement of existing imaging sensors is likely to generate more immediate impacts."
Image sensors, which converts light into electrical signals, are composed of millions of pixels on a single chip. The challenge is how to combine and miniaturize multifunctional components as part of the sensor.
In their own work, the researchers detailed a promising approach to detect multiple-band spectra by fabricating an on-chip spectrometer. They deposited photonic crystal filters made up of silicon directly on top of the pixels to create complex interactions between incident light and the sensor.
The pixels beneath the films record the distribution of light energy, from which light spectral information can be inferred. The device—less than a hundredth of a square inch in size—is programmable to meet various dynamic ranges, resolution levels, and almost any spectral regime from visible to infrared.
The researchers built a component that detects angular information to measure depth and construct 3D shapes at subcellular scales. Their work was inspired by directional hearing sensors found in animals, like geckos, whose heads are too small to determine where sound is coming from in the same way humans and other animals can. Instead, they use coupled eardrums to measure the direction of sound within a size that is orders of magnitude smaller than the corresponding acoustic wavelength.
Similarly, pairs of silicon nanowires were constructed as resonators to support optical resonance. The optical energy stored in two resonators is sensitive to the incident angle. The wire closest to the light sends the strongest current. By comparing the strongest and weakest currents from both wires, the angle of the incoming light waves can be determined.
Millions of these nanowires can be placed on a 1-square-millimeter chip. The research could support advances in lensless cameras, augmented reality, and robotic vision.
The article "Multimodal light-sensing pixel arrays" is authored by Yurui Qu, Soongyu Yi, Lan Yang, and Zongfu Yu. The article will appear in Applied Physics Letters on July 26, 2022.
More information: Multimodal light-sensing pixel arrays, Applied Physics Letters (2022). DOI: 10.1063/5.0090138
Journal information: Applied Physics Letters
Provided by American Institute of Physics