Social robots see smell

Aug 13, 2013 by Angela Herring
Bioengineering graduate student Ryan Myers built a sophisticated printer that can create microelectronics small enough to integrate with living cells. Professor Joseph Ayers is using these devices to give his robots a sense of smell.

"The thing that's been missing in robotics is a sense of smell," said biology professor Joseph Ayers.

For more than four decades, he has been working to develop robots that do not rely on algorithms or external controllers. Instead, they incorporate electronic nervous systems that take in from the environment and spit out autonomous behaviors. For example, his team's robo-lobsters are designed to seek out without following a predetermined course.

"Now people want robots to do ," said Ayers, noting that social insect colonies are the perfect model. "If you're doing large field explorations for mines, you want to have 20 or 30 robots out there." In order to get robots to cooperate with each other, he needs them to act like ants or bees or termites.

Bees waggle their behinds to communicate. Ants use almost two dozen scent glands, depositing a trail of "stinks" as they go about their business. It's this behavior that Ayers wants to mimic in his next generation of biomimetic robots.

To do so, he needs electronic devices that can sense chemical inputs, such as explosives. His idea is to integrate various microelectronic sensors that can interface with living cells. For example, a bacterial cell programmed to bind odorants in the environment may elicit a conformational change; that change may translate to an influx of , which are detected by a second cell that is programmed to generate light when bound to calcium. In this way, Ayers said, "you can see smell."

That output would then trigger microelectronic actuators that tell the robot to perform a particular action, such as moving toward or away from the stimulus.

But in order for any of this to play out, somebody needs to build these futuristic devices.

Enter bioengineering graduate student Ryan Myers, who built one of the world's only e-jet printers for Ayers' lab. He learned the nearly artisanal craft from Andrew Alleyne, a professor of engineering at the University of Illinois who perfected the technology. Myers' work earned him the interdisciplinary research award at the Research, Innovation, Scholarship, and Entrepreneurship expo earlier this year.

According to Ayers, "inkjet printing is the industry standard for organic electronics." This state-of-the-art technology is already paving the way for a new industry of inexpensive, versatile electronics, such as the curved television that debuted earlier this month.

The problem, at least for Ayers' lab, is that inkjet printers can only deposit droplets 30 microns or larger. While that might seem sufficiently teeny to the rest of us, it's not small enough for Ayers, who needs electronics features that are smaller than a living cell.

That's where the "e," for electrohydrodynamic, comes into the picture. In the case of "traditional" inkjets, a droplet is deposited onto a surface through backpressure alone. This means that some of the ink spreads out when it lands. E-jet printers incorporate a voltage potential between the printer head and the surface, as well as a small vacuum force on the other side. When the ink drops from the printer head, it is both pushed and pulled to the exact spot for which it's intended. The technology allows them to print droplets as small as 250 nanometers.

At a fraction of the diameter of a living cell, "we can print many features per cell instead of many cells per feature," said Ayers. That is, they can now produce microelectronics with high enough resolution to integrate with biological systems.

The research team is now hard at work printing biocompatible photodiodes, nitric oxide sensors, and photosensors to integrate into their robo-lobster and rob-lamprey projects. It's just the next step in Ayers' goal to create a "social-robot."

Explore further: Robots recognize humans in disaster environments

add to favorites email to friend print save as pdf

Related Stories

RoboBees get smart in pollen pursuit

Jun 10, 2013

(Phys.org) —When a scout honeybee returns to the hive, she performs a "waggle dance," looping and shaking her rear end in particular patterns to direct her comrades toward the jackpot of nectar and pollen ...

New underwater robot swims and senses like a fish

Jul 18, 2013

In recent years, robotic underwater vehicles have become more common in a variety of industrial and civil sectors. They are used extensively by the scientific community to study the ocean. For example, underwater ...

3Qs: The 3-D printing of tomorrow

Mar 15, 2013

Ahmed Busnaina, the William Lincoln Smith Professor and director of the NSF Nanoscale Science and Engineering Center for High-rate Nanomanufacturing at Northeastern, has developed a method called directed ...

Recommended for you

Robots recognize humans in disaster environments

Oct 21, 2014

Through a computational algorithm, a team of researchers from the University of Guadalajara (UDG) in Mexico, developed a neural network that allows a small robot to detect different patterns, such as images, ...

Japan toymaker unveils tiny talking, singing humanoid

Oct 15, 2014

Japanese toymaker Tomy on Wednesday unveiled a multi-talented humanoid robot, named "Robi jr.," which can converse using some 1,000 phrases and belt out about 50 songs, as well as move its limbs and head.

User comments : 0