Engineers Develop Undetectable Means Of Measuring Speed, Motion

Mar 30, 2005

Research aimed at teaching robots to "see" may soon make it possible to bag speeding motorists, track enemy planes, and automatically safeguard the nation’s borders and resources without any chance of detection.
A University of Florida engineering researcher is developing a method to determine speed and other characteristics of a moving object using computer algorithms, or instructions, that rely on data from standard visual cameras rather than radar or lasers. The technique has the potential to render current detection systems in so-called "fuzz busters" and some military technologies useless. That’s because, instead of painting a target with radar waves or laser beams, a camera merely needs to capture an image or series of images from the target.

“If it can view the object moving, that’s all it needs. The computer figures out everything else,” said Warren Dixon, a UF assistant professor of mechanical and aerospace engineering. “We’re trying to use both regular and infrared cameras, so night or adverse weather conditions don’t present a problem.”

Dixon’s most recent co-authored article on the research appears in the March issue of the journal Automatica. Related articles, also co-authored by Dixon, are scheduled to appear shortly in the journal Transactions on Robotics and Automation.

Achieving computerized speed and motion detection requires overcoming several challenges. One is figuring out how to get a computer to understand the surrounding environment by interpreting images recorded by a video or still camera.

“The information from a camera is just a flat-screen, two-dimensional image,” Dixon said. “The challenge is figuring out the mathematics of how do you take two images and understand how things are moving in our three-dimensional world.”

People and animals can perceive depth because their brains combine each eye’s snapshots. Two cameras can also achieve stereo vision, but computers can make sense of it only if they know the exact position of each camera. That allows them to triangulate the target and learn its position relative to the camera. Part of Dixon’s achievement is developing the underlying mathematics and software to circumvent this requirement.

“With my work, you don’t need to know that specific location information,” he said. “You could have one camera taking an image from an airplane and another mounted on a car taking a picture of the same image -- and not know how the airplane and car are related to each other -- and through this new mathematics you can understand how they’re related to the target.”

The technology has law enforcement and military applications.

Police in moving or parked squad cars could use the computer-camera systems much as they do radar and laser guns to track and ticket suspected speeders. The target would have to be within the line of sight, with the range varying according to the power of the lenses in the camera. Dixon said the UF engineers have not built such a system, but “any camera with the right software could be used,” and a prototype could be built within a year.

Soldiers, meanwhile, could mount the cameras on airborne drones or truck convoys and set them to look for and automatically report potentially hostile objects moving toward the convoys – again, without any fear of giving away the convoys’ locations.

Robotic drones or remote camera-based monitoring posts outfitted with the technology also could be used for applications ranging from private security in warehouses and shopping centers to continuous remote monitoring of borders to protecting water supply reservoirs.

In addition to the robotic applications, the technique is being refined for a project led by Andy Kurdila, a UF professor of mechanical and aerospace engineering, to provide vision systems for tiny airborne surveillance drones called micro air vehicles.

The goal of that five-year project, which is jointly funded by a $5 million grant from Eglin Air Force Base in Florida and by the Air Force Office of Scientific Research and involves several UF faculty members, is to give the drones the ability to fly without the assistance of a remote human operator. Instead, they would base navigational decisions solely on what onboard cameras view in the terrain as they fly, mimicking a human pilot.

Explore further: Explainer: How to solve a jewel heist (and why it takes so long)

Related Stories

Engineering mechanical parts for space activity

May 20, 2015

For more than four decades, an unremarkable building in an industrial estate on the edge of Warrington, UK, has played a crucial role in the success of most European space missions with moving parts – which ...

Sandia helps small security company thwart thieves

May 19, 2015

At a motorcycle shop on a busy city street, crooks devised an elaborate scheme to steal from the storage yard. They jumped the fence and unpacked some newly arrived bikes from crates. They used the crates ...

Recommended for you

Top UK scientists warn against EU exit

7 hours ago

A group of leading British scientists including Nobel-winning geneticist Paul Nurse warned leaving the European Union could threaten research funding, in a letter published in The Times newspaper on Friday.

How we discovered the three revolutions of American pop

8 hours ago

Dr Matthias Mauch discusses his recent scientific analysis of the "fossil record" of the Billboard charts prompted widespread attention, particularly the findings about the three musical "revolutions" that shaped the musical la ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.