Flight-tests autonomous multi-target, multi-user tracking capability

Aug 17, 2011 By Daniel Parry
Graphic depiction of the network sensing concept.

The Naval Research Laboratory and the Space Dynamics Laboratory (SDL) through the support of the Office of Naval Research (ONR), has shown an autonomous multi-sensor motion-tracking and interrogation system that reduces the workload for analysts by automatically finding moving objects, then presenting high-resolution images of those objects with no human input.

Intelligence, surveillance and reconnaissance (ISR) assets in the field generate vast amounts of data that can overwhelm human operators and can severely limit the ability of an analyst to generate intelligence reports in operationally relevant timeframes. This multi-user tracking capability enables the system to manage collection of imagery without continuous monitoring by a ground or airborne operator, thus requiring fewer personnel and freeing up operational assets.

"These tests display how a single imaging sensor can be used to provide imagery of multiple tracked objects," said Dr. Brian Daniel, research physicist, NRL ISR Systems and Processing Section. "A job typically requiring multiple sensors."

During , March 2011, multiple real-time tracks generated by a wide-area persistent surveillance sensor (WAPSS) were autonomously cross-cued to a high-resolution narrow field-of-view (NFOV) interrogation sensor via an airborne network. Both sensors were networked by the high-speed Tactical Reachback Extended Communications, TREC, data-link provided by the NRL Information Technology Division, Satellite and Wireless Technology Branch.

"The demonstration was a complete success," noted Dr. Michael Duncan, ONR program manager. "Not only did the network sensing demonstration achieve simultaneous real-time tracking, sensor cross cueing and inspection of multiple vehicle-sized objects, but we also showed an ability to follow smaller human-sized objects under specialized conditions."

The network sensing demonstration utilized built under other ONR sponsored programs. The interrogation sensor was the precision, jitter-stabilized EyePod developed under the Fusion, Exploitation, Algorithm, and Targeting High-Altitude Reconnaissance (FEATHAR) program. EyePod is a dual-band visible-near infrared and long-wave infrared sensor mounted inside a nine-inch gimbal pod assembly designed for small UAV platforms. The mid-wave infrared nighttime WAPSS (N-WAPSS) was chosen as the wide-area sensor, and has a 16 mega-pixel, large format camera that captures single frames at four hertz (cycles per second) and has a step-stare capability with a one hertz refresh rate.

Using precision geo-projection of the N-WAPSS imagery, all moving vehicle-size objects in the FOV were tracked in real-time. The tracks were converted to geodetic coordinates and sent via an air-based network to a cue manager system. The cue manager autonomously tasked EyePod to interrogate all selected tracks for target classification and identification.

Explore further: MIT team's wireless Vital-Radio could follow breathing, heart rate at home

Related Stories

ONR-guided tech tracks what's inside ships

Apr 01, 2010

The Office of Naval Research (ONR) is funding emerging technology that will allow wireless surveillance not only of ships and aircraft, but also the tracking of people and high value assets inside the ships.

Proxy Aviation Systems Unveils SkyWatcher

Jun 28, 2005

Proxy Aviation Systems recently unveiled SkyWatcher, a long endurance, low and medium altitude, multi-payload unmanned aircraft system at the Association for Unmanned Vehicle Systems International's (AUVSI) demonstration ...

Tampa Bay Becomes "Smart Bay" With Well-Placed Sensors

Aug 17, 2005

A demonstration in July in which scientists and engineers from the University of South Florida placed sophisticated, small, rugged sensors at strategic points in Tampa Bay and downloaded data from them wirelessly illustrated ...

Recommended for you

Team develops faster, higher quality 3-D camera

Apr 24, 2015

When Microsoft released the Kinect for Xbox in November 2010, it transformed the video game industry. The most inexpensive 3-D camera to date, the Kinect bypassed the need for joysticks and controllers by ...

Researchers finding applications for tough spinel ceramic

Apr 24, 2015

Imagine a glass window that's tough like armor, a camera lens that doesn't get scratched in a sand storm, or a smart phone that doesn't break when dropped. Except it's not glass, it's a special ceramic called ...

Classroom acoustics for architects

Apr 23, 2015

The Acoustical Society of America (ASA) has published a free online booklet for architects to aid in the application of ANSI/ASA S12.60-2010/Part 1-American National Standard Acoustical Performance Criteria, Design Requirements, ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.