New technology could make TV more exciting

Feb 02, 2005

Live TV outside broadcasts that combine real action and computer-generated images could become possible for the first time, thanks to camera navigation technology now under development.
The work is opening up the prospect of outdoor sporting, musical or other TV coverage that blends the excitement of being live with the spectacular visual impact that computer graphics can create. It can also be applied at the consumer level, e.g. to enable interior design ideas to be visualised by adding virtual furniture to the view of a room provided by a hand-held camera as it moves.

The system is able to work out in real-time where a camera is and how it is moving, simultaneously constructing a detailed visual map of its surroundings. This enables computer graphics to be overlaid accurately onto live pictures as soon as they are produced. Previously the blending of live action and computer-generated images has only been possible in controlled studio environments.

Harnessing techniques from mathematics, computing and engineering, the new system is being developed at Oxford University with funding from the Engineering and Physical Sciences Research Council (EPSRC). The project aims to extend the capabilities of a prototype system developed by the same team, also with EPSRC funding.

Dr Ian Reid and Dr Andrew Davison of Oxford University's Department of Engineering Science are leading the project. Andrew says: "This localisation and mapping technology turns a camera into a flexible, real-time position sensor. It has all kinds of potential applications."

The system comprises a mobile video camera connected to a laptop computer, which analyses the images it receives using software developed by the researchers. As the camera moves, the system picks out landmarks as reference points and makes a map of their 3D locations against which to measure its position. The challenge is to estimate accurately the camera's position and the layout of its surroundings at the same time - a task known as Simultaneous Localisation and Mapping (SLAM).

As well as TV and video applications, the technology under development could provide low-cost, high-performance navigation for domestic robots. It could also be incorporated into video games or wearable computing, e.g. for use in dangerous environments, where it could confirm the wearer's location and allow relevant guidance to be overlaid onto their view of surroundings.

Source: Engineering and Physical Sciences Research Council

Explore further: Social Security spent $300M on 'IT boondoggle'

add to favorites email to friend print save as pdf

Related Stories

Medical advances turn science fiction into science fact

Jul 18, 2014

Exoskeletons helping the paralysed to walk, tiny maggot-inspired devices gnawing at brain tumours, machines working tirelessly as hospital helpers: in many respects, the future of medicine is already here.

ESA's spaceplane set for flight

Jul 16, 2014

(Phys.org) —All eyes are on ESA's spaceplane to showcase reentry technologies after its unconventional launch on a Vega rocket this November.

Lighting for photography with drones

Jul 11, 2014

Lighting is crucial to the art of photography. But lights are cumbersome and time-consuming to set up, and outside the studio, it can be prohibitively difficult to position them where, ideally, they ought ...

Recommended for you

Hoverbike drone project for air transport takes off

2 hours ago

What happens when you cross a helicopter with a motorbike? The crew at Malloy Aeronautics has been focused on a viable answer and has launched a crowdfunding campaign to support its Hoverbike project, "The ...

Study shows role of media in sharing life events

3 hours ago

To share is human. And the means to share personal news—good and bad—have exploded over the last decade, particularly social media and texting. But until now, all research about what is known as "social sharing," or the ...

User comments : 0