Next-gen computing: Memristor chips that see patterns over pixels

May 22, 2017, University of Michigan

Inspired by how mammals see, a new "memristor" computer circuit prototype at the University of Michigan has the potential to process complex data, such as images and video orders of magnitude, faster and with much less power than today's most advanced systems.

Faster image processing could have big implications for autonomous systems such as self-driving cars, says Wei Lu, U-M professor of electrical engineering and science. Lu is lead author of a paper on the work published in the current issue of Nature Nanotechnology.

Lu's next-generation computer components use pattern recognition to shortcut the energy-intensive process conventional systems use to dissect images. In this new work, he and his colleagues demonstrate an algorithm that relies on a technique called "sparse coding" to coax their 32-by-32 array of memristors to efficiently analyze and recreate several photos.

Memristors are electrical resistors with memory—advanced electronic devices that regulate current based on the history of the voltages applied to them. They can store and process data simultaneously, which makes them a lot more efficient than traditional systems. In a conventional computer, logic and memory functions are located at different parts of the circuit.

"The tasks we ask of today's computers have grown in complexity," Lu said. "In this 'big data' era, computers require costly, constant and slow communications between their processor and memory to retrieve large amounts data. This makes them large, expensive and power-hungry."

But like neural networks in a biological brain, networks of memristors can perform many operations at the same time, without having to move data around. As a result, they could enable new platforms that process a vast number of signals in parallel and are capable of advanced . Memristors are good candidates for deep , a branch of machine learning, which trains computers to execute processes without being explicitly programmed to do so.

"We need our next-generation electronics to be able to quickly process complex data in a dynamic environment. You can't just write a program to do that. Sometimes you don't even have a pre-defined task," Lu said. "To make our systems smarter, we need to find ways for them to process a lot of data more efficiently. Our approach to accomplish that is inspired by neuroscience."

A mammal's brain is able to generate sweeping, split-second impressions of what the eyes take in. One reason is because they can quickly recognize different arrangements of shapes. Humans do this using only a limited number of neurons that become active, Lu says. Both neuroscientists and computer scientists call the process "sparse coding."

"When we take a look at a chair we will recognize it because its characteristics correspond to our stored mental picture of a chair," Lu said. "Although not all chairs are the same and some may differ from a mental prototype that serves as a standard, each chair retains some of the key characteristics necessary for easy recognition. Basically, the object is correctly recognized the moment it is properly classified—when 'stored' in the appropriate category in our heads."

Similarly, Lu's electronic system is designed to detect the patterns very efficiently—and to use as few features as possible to describe the original input.

In our brains, different neurons recognize different patterns, Lu says.

"When we see an image, the neurons that recognize it will become more active," he said. "The neurons will also compete with each other to naturally create an efficient representation. We're implementing this approach in our electronic system."

The researchers trained their system to learn a "dictionary" of images. Trained on a set of grayscale image patterns, their network was able to reconstruct images of famous paintings and photos and other test patterns.

If their system can be scaled up, they expect to be able to process and analyze video in real time in a compact system that can be directly integrated with sensors or cameras.

Explore further: Image-processing 1,000 times faster is goal of new $5M contract

More information: Sparse coding with memristor networks, Nature Nanotechnology (2017).

Related Stories

Cat brain: A step toward the electronic equivalent

April 14, 2010

A cat can recognize a face faster and more efficiently than a supercomputer. That's one reason a feline brain is the model for a biologically-inspired computer project involving the University of Michigan.

Computer learns to recognize sounds by watching video

December 1, 2016

In recent years, computers have gotten remarkably good at recognizing speech and images: Think of the dictation software on most cellphones, or the algorithms that automatically identify people in photos posted to Facebook.

Recommended for you

Plasmons triggered in nanotube quantum wells

March 16, 2018

A novel quantum effect observed in a carbon nanotube film could lead to the development of unique lasers and other optoelectronic devices, according to scientists at Rice University and Tokyo Metropolitan University.

Zero field switching (ZFS) effect in a nanomagnetic device

March 16, 2018

An unexpected phenomenon known as zero field switching (ZFS) could lead to smaller, lower-power memory and computing devices than presently possible. The image shows a layering of platinum (Pt), tungsten (W), and a cobalt-iron-boron ...

Imaging technique pulls plasmon data together

March 16, 2018

Rice University scientists have developed a novel technique to view a field of plasmonic nanoparticles simultaneously to learn how their differences change their reactivity.


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.