Enhanced video quality despite poor network conditions

Enhanced video quality despite poor network conditions
Figure 1. Image quality before (left) and after (right) the technology application. Credit: The Korea Advanced Institute of Science and Technology (KAIST)

Professor Jinwoo Shin and Professor Dongsu Han from the School of Electrical Engineering developed neural adaptive content-aware internet video delivery. This technology is a novel method that combines adaptive streaming over HTTP, the video transmission system adopted by YouTube and Netflix, with a deep learning model.

This technology is expected to create an internet environment where users can enjoy watching 4K and AV/VR videos with high-quality and high-definition (HD) videos even with weak internet connections.

Thanks to streaming services, internet video has experienced remarkable growth; nevertheless, users often suffer from low video quality due to unfavorable network conditions. Currently, existing adaptive streaming systems adjust the quality of the video in , accommodating the continuously changing internet bandwidth. Various algorithms are being researched for adaptive streaming systems, but there is an inherent limitation; that is, high-quality videos cannot be streamed in poor network environments regardless of which algorithm is used.

By incorporating super-resolution in adaptive streaming, the team overcame the limit of existing content distribution networks, of which their quality relies too much on the bandwidth. In the conventional method, the server that provides the video splits a video into certain lengths of time in advance. But the novel system introduced by the team allows the downloading of neural network segments. To facilitate this method, the video server needs to provide for each video segment as well as sizes of Deep Neural Networks (DNN) according to the specifications of the user's computing capacity.

Enhanced video quality despite poor network conditions
Figure 2. The technology Concept. Credit: The Korea Advanced Institute of Science and Technology (KAIST)

The largest neural network size is two megabytes, which is considerably smaller than video. When downloading the neural network from the user's video player, it is split into several segments. Even its partial download is sufficient for a slightly comprised super-resolution.

While playing the video, the system converts the low quality video to a high-quality version by employing super-resolution based on deep convolution neural networks (CNN). The entire process is done in real time, and users can enjoy the video.

Even with a 17 percent smaller bandwidth, the system can provide the Quality of Experience equivalent to the latest adaptive streaming service. At a given internet bandwidth, it can provide 43 percent higher average QoE than the latest service.

Enhanced video quality despite poor network conditions
Figure 3. A transition from low-quality to high quality video after video transmission from the video server. Credit: The Korea Advanced Institute of Science and Technology (KAIST)

Using a deep learning method allows this system to achieve a higher level of compression than the existing video compression methods. Their technology was recognized as a next-generation video system that applies super-resolution based on a deep convolution neural network to online videos.

Professor Han said, "So far, it has only been implemented on desktops, but we will further develop applications that work in mobile devices as well. This has been applied to the same video transmission systems used by streaming channels such as YouTube and Netflix, and thus shows good signs for practicability."

More information: Neural Adaptive Content-aware Internet Video Delivery. www.usenix.org/conference/osdi18/presentation/yeo

Citation: Enhanced video quality despite poor network conditions (2019, January 22) retrieved 15 July 2024 from https://phys.org/news/2019-01-video-quality-poor-network-conditions.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Smart prioritization of visually important data to improve video streaming


Feedback to editors