Dr Stamos Katsigiannis stamos.katsigiannis@durham.ac.uk
Associate Professor
A machine learning driven solution to the problem of perceptual video quality metrics
Katsigiannis, Stamos; Rabah, Hassan; Ramzan, Naeem
Authors
Hassan Rabah
Naeem Ramzan
Contributors
Muhammad Zeeshan Shakir
Editor
Naeem Ramzan
Editor
Abstract
The advent of high-speed internet connections, advanced video coding algorithms, and consumer-grade computers with high computational capabilities has led videostreaming-over-the-internet to make up the majority of network traffic. This effect has led to a continuously expanding video streaming industry that seeks to offer enhanced quality-of-experience (QoE) to its users at the lowest cost possible. Video streaming services are now able to adapt to the hardware and network restrictions that each user faces and thus provide the best experience possible under those restrictions. The most common way to adapt to network bandwidth restrictions is to offer a video stream at the highest possible visual quality, for the maximum achievable bitrate under the network connection in use. This is achieved by storing various pre-encoded versions of the video content with different bitrate and visual quality settings. Visual quality is measured by means of objective quality metrics, such as the Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Visual Information Fidelity (VIF), and others, which can be easily computed analytically. Nevertheless, it is widely accepted that although these metrics provide an accurate estimate of the statistical quality degradation, they do not reflect the viewer’s perception of visual quality accurately. As a result, the acquisition of user ratings in the form of Mean Opinion Scores (MOS) remains the most accurate depiction of human-perceived video quality, albeit very costly and time consuming, and thus cannot be practically employed by video streaming providers that have hundreds or thousands of videos in their catalogues. A recent very promising approach for addressing this limitation is the use of machine learning techniques in order to train models that represent human video quality perception more accurately. To this end, regression techniques are used in order to map objective quality metrics to human video quality ratings, acquired for a large number of diverse video sequences. Results have been very promising, with approaches like the Video Multimethod Assessment Fusion (VMAF) metric achieving higher correlations to useracquired MOS ratings compared to traditional widely used objective quality metrics.
Citation
Katsigiannis, S., Rabah, H., & Ramzan, N. (2020). A machine learning driven solution to the problem of perceptual video quality metrics. In M. Z. Shakir, & N. Ramzan (Eds.), AI for Emerging Verticals; Human-robot computing, sensing and networking. IET
Online Publication Date | Dec 15, 2020 |
---|---|
Publication Date | 2020 |
Deposit Date | Dec 15, 2020 |
Publicly Available Date | Jan 13, 2021 |
Publisher | IET |
Book Title | AI for Emerging Verticals; Human-robot computing, sensing and networking. |
Public URL | https://durham-repository.worktribe.com/output/1626378 |
Publisher URL | https://shop.theiet.org/ai-for-emerging-verticals |
Files
Accepted Book Chapter
(347 Kb)
PDF
You might also like
Toward Automatic Tutoring of Math Word Problems in Intelligent Tutoring Systems
(2023)
Journal Article
Downloadable Citations
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search