Smarter video searching and indexing

A new framework being created by a PhD student and academics from the University of Lincoln, UK, will enable people to search for videos on the internet, using images rather than text.

Typing in text to find a film clip on YouTube often results in diverse (and sometimes unrelated) videos being suggested.

This problem could soon be resolved with the advent of smarter - engines that are able to pick and choose the most relevant videos by analysing a tiny fraction of video frames.

Research to create a quick and easy that is able to discover semantic similarities between videos without using text-based tags is being carried out at the University of Lincoln, UK.

The volume of video data is rapidly increasing with more than 4 billion hours of video being watched on YouTube each month. The majority of available exists in compressed format and the first step towards effective video retrieval is to extract features from these compressed videos.

School of Computer Science PhD student Saddam Bekhet, along with Dr Amr Ahmed and Professor Andrew Hunter, has produced a paper on recent work which suggests a framework towards real-time video matching.

Saddam said: "Everyone uses search engines but currently you are only able to search by text even to search for a video clip, thus some results are far removed from what you were looking for. With the huge volume of data, a smarter video analyser is required to associate semantic tags to the uploaded videos, allowing more efficient indexing and search (including the contents of the video). Being able to enhance the underlying search mechanism (or even input a visual query) would really enhance the likes of YouTube."

Saddam's framework relies upon finding similarities between videos using tiny frames instead of using the full-size video frames. Such tiny frames are easily extracted from a compressed video in real-time and able to fully represent video content, without wasting more time in decompressing the video to perform complex computer algorithms.

"I want to discover the semantic similarity between videos using the content only," he explained. "I adapted some new techniques and found that tiny representative frames could be used to discover similarities. The next stage is to build an effective framework."

The research follows on from work carried out to provide a framework for automated video analysis and annotation, by Lincoln's Digital Contents Analysis, Production and Interaction (DCAPI) Group.

The paper 'Video Matching Using DC-image and Local Features' was awarded the Best Student Paper Award at the International Conference of Signal and Image Engineering - part of the 2013 World Congress on Engineering. Organised by the International Association of Engineers (IAENG), the conference focuses on the frontier topics in the theoretical and applied engineering and computer science subjects.

Citation: Smarter video searching and indexing (2013, October 14) retrieved 5 May 2024 from https://phys.org/news/2013-10-smarter-video-indexing.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

New system can improve video-sharing Web sites like YouTube

0 shares

Feedback to editors