Loading…

Convolutional Neural Networks for Video Quality Assessment

Video Quality Assessment (VQA) is a very challenging task due to its highly subjective nature. Moreover, many factors influence VQA. Compression of video content, while necessary for minimising transmission and storage requirements, introduces distortions which can have detrimental effects on the pe...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2018-09
Main Authors: Giannopoulos, Michalis, Tsagkatakis, Grigorios, Blasi, Saverio, Toutounchi, Farzad, Mouchtaris, Athanasios, Tsakalides, Panagiotis, Mrak, Marta, Izquierdo, Ebroul
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Giannopoulos, Michalis
Tsagkatakis, Grigorios
Blasi, Saverio
Toutounchi, Farzad
Mouchtaris, Athanasios
Tsakalides, Panagiotis
Mrak, Marta
Izquierdo, Ebroul
description Video Quality Assessment (VQA) is a very challenging task due to its highly subjective nature. Moreover, many factors influence VQA. Compression of video content, while necessary for minimising transmission and storage requirements, introduces distortions which can have detrimental effects on the perceived quality. Especially when dealing with modern video coding standards, it is extremely difficult to model the effects of compression due to the unpredictability of encoding on different content types. Moreover, transmission also introduces delays and other distortion types which affect the perceived quality. Therefore, it would be highly beneficial to accurately predict the perceived quality of video to be distributed over modern content distribution platforms, so that specific actions could be undertaken to maximise the Quality of Experience (QoE) of the users. Traditional VQA techniques based on feature extraction and modelling may not be sufficiently accurate. In this paper, a novel Deep Learning (DL) framework is introduced for effectively predicting VQA of video content delivery mechanisms based on end-to-end feature learning. The proposed framework is based on Convolutional Neural Networks, taking into account compression distortion as well as transmission delays. Training and evaluation of the proposed framework are performed on a user annotated VQA dataset specifically created to undertake this work. The experiments show that the proposed methods can lead to high accuracy of the quality estimation, showcasing the potential of using DL in complex VQA scenarios.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2112962371</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2112962371</sourcerecordid><originalsourceid>FETCH-proquest_journals_21129623713</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcs7PK8vPKS3JzM9LzFHwSy0tAlMl5flF2cUKaflFCmGZKan5CoGliTmZJZUKjsXFqcXFual5JTwMrGmJOcWpvFCam0HZzTXE2UO3oCi_sDS1uCQ-K7-0CGhqcbyRoaGRpZmRsbmhMXGqAPy2N0E</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2112962371</pqid></control><display><type>article</type><title>Convolutional Neural Networks for Video Quality Assessment</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Giannopoulos, Michalis ; Tsagkatakis, Grigorios ; Blasi, Saverio ; Toutounchi, Farzad ; Mouchtaris, Athanasios ; Tsakalides, Panagiotis ; Mrak, Marta ; Izquierdo, Ebroul</creator><creatorcontrib>Giannopoulos, Michalis ; Tsagkatakis, Grigorios ; Blasi, Saverio ; Toutounchi, Farzad ; Mouchtaris, Athanasios ; Tsakalides, Panagiotis ; Mrak, Marta ; Izquierdo, Ebroul</creatorcontrib><description>Video Quality Assessment (VQA) is a very challenging task due to its highly subjective nature. Moreover, many factors influence VQA. Compression of video content, while necessary for minimising transmission and storage requirements, introduces distortions which can have detrimental effects on the perceived quality. Especially when dealing with modern video coding standards, it is extremely difficult to model the effects of compression due to the unpredictability of encoding on different content types. Moreover, transmission also introduces delays and other distortion types which affect the perceived quality. Therefore, it would be highly beneficial to accurately predict the perceived quality of video to be distributed over modern content distribution platforms, so that specific actions could be undertaken to maximise the Quality of Experience (QoE) of the users. Traditional VQA techniques based on feature extraction and modelling may not be sufficiently accurate. In this paper, a novel Deep Learning (DL) framework is introduced for effectively predicting VQA of video content delivery mechanisms based on end-to-end feature learning. The proposed framework is based on Convolutional Neural Networks, taking into account compression distortion as well as transmission delays. Training and evaluation of the proposed framework are performed on a user annotated VQA dataset specifically created to undertake this work. The experiments show that the proposed methods can lead to high accuracy of the quality estimation, showcasing the potential of using DL in complex VQA scenarios.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Artificial neural networks ; Coding standards ; Compression tests ; Distortion ; Feature extraction ; Machine learning ; Mathematical models ; Neural networks ; Quality ; Quality assessment ; User satisfaction ; Video compression ; Video transmission</subject><ispartof>arXiv.org, 2018-09</ispartof><rights>2018. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2112962371?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Giannopoulos, Michalis</creatorcontrib><creatorcontrib>Tsagkatakis, Grigorios</creatorcontrib><creatorcontrib>Blasi, Saverio</creatorcontrib><creatorcontrib>Toutounchi, Farzad</creatorcontrib><creatorcontrib>Mouchtaris, Athanasios</creatorcontrib><creatorcontrib>Tsakalides, Panagiotis</creatorcontrib><creatorcontrib>Mrak, Marta</creatorcontrib><creatorcontrib>Izquierdo, Ebroul</creatorcontrib><title>Convolutional Neural Networks for Video Quality Assessment</title><title>arXiv.org</title><description>Video Quality Assessment (VQA) is a very challenging task due to its highly subjective nature. Moreover, many factors influence VQA. Compression of video content, while necessary for minimising transmission and storage requirements, introduces distortions which can have detrimental effects on the perceived quality. Especially when dealing with modern video coding standards, it is extremely difficult to model the effects of compression due to the unpredictability of encoding on different content types. Moreover, transmission also introduces delays and other distortion types which affect the perceived quality. Therefore, it would be highly beneficial to accurately predict the perceived quality of video to be distributed over modern content distribution platforms, so that specific actions could be undertaken to maximise the Quality of Experience (QoE) of the users. Traditional VQA techniques based on feature extraction and modelling may not be sufficiently accurate. In this paper, a novel Deep Learning (DL) framework is introduced for effectively predicting VQA of video content delivery mechanisms based on end-to-end feature learning. The proposed framework is based on Convolutional Neural Networks, taking into account compression distortion as well as transmission delays. Training and evaluation of the proposed framework are performed on a user annotated VQA dataset specifically created to undertake this work. The experiments show that the proposed methods can lead to high accuracy of the quality estimation, showcasing the potential of using DL in complex VQA scenarios.</description><subject>Artificial neural networks</subject><subject>Coding standards</subject><subject>Compression tests</subject><subject>Distortion</subject><subject>Feature extraction</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>Neural networks</subject><subject>Quality</subject><subject>Quality assessment</subject><subject>User satisfaction</subject><subject>Video compression</subject><subject>Video transmission</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mSwcs7PK8vPKS3JzM9LzFHwSy0tAlMl5flF2cUKaflFCmGZKan5CoGliTmZJZUKjsXFqcXFual5JTwMrGmJOcWpvFCam0HZzTXE2UO3oCi_sDS1uCQ-K7-0CGhqcbyRoaGRpZmRsbmhMXGqAPy2N0E</recordid><startdate>20180926</startdate><enddate>20180926</enddate><creator>Giannopoulos, Michalis</creator><creator>Tsagkatakis, Grigorios</creator><creator>Blasi, Saverio</creator><creator>Toutounchi, Farzad</creator><creator>Mouchtaris, Athanasios</creator><creator>Tsakalides, Panagiotis</creator><creator>Mrak, Marta</creator><creator>Izquierdo, Ebroul</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20180926</creationdate><title>Convolutional Neural Networks for Video Quality Assessment</title><author>Giannopoulos, Michalis ; Tsagkatakis, Grigorios ; Blasi, Saverio ; Toutounchi, Farzad ; Mouchtaris, Athanasios ; Tsakalides, Panagiotis ; Mrak, Marta ; Izquierdo, Ebroul</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_21129623713</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Artificial neural networks</topic><topic>Coding standards</topic><topic>Compression tests</topic><topic>Distortion</topic><topic>Feature extraction</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>Neural networks</topic><topic>Quality</topic><topic>Quality assessment</topic><topic>User satisfaction</topic><topic>Video compression</topic><topic>Video transmission</topic><toplevel>online_resources</toplevel><creatorcontrib>Giannopoulos, Michalis</creatorcontrib><creatorcontrib>Tsagkatakis, Grigorios</creatorcontrib><creatorcontrib>Blasi, Saverio</creatorcontrib><creatorcontrib>Toutounchi, Farzad</creatorcontrib><creatorcontrib>Mouchtaris, Athanasios</creatorcontrib><creatorcontrib>Tsakalides, Panagiotis</creatorcontrib><creatorcontrib>Mrak, Marta</creatorcontrib><creatorcontrib>Izquierdo, Ebroul</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Giannopoulos, Michalis</au><au>Tsagkatakis, Grigorios</au><au>Blasi, Saverio</au><au>Toutounchi, Farzad</au><au>Mouchtaris, Athanasios</au><au>Tsakalides, Panagiotis</au><au>Mrak, Marta</au><au>Izquierdo, Ebroul</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Convolutional Neural Networks for Video Quality Assessment</atitle><jtitle>arXiv.org</jtitle><date>2018-09-26</date><risdate>2018</risdate><eissn>2331-8422</eissn><abstract>Video Quality Assessment (VQA) is a very challenging task due to its highly subjective nature. Moreover, many factors influence VQA. Compression of video content, while necessary for minimising transmission and storage requirements, introduces distortions which can have detrimental effects on the perceived quality. Especially when dealing with modern video coding standards, it is extremely difficult to model the effects of compression due to the unpredictability of encoding on different content types. Moreover, transmission also introduces delays and other distortion types which affect the perceived quality. Therefore, it would be highly beneficial to accurately predict the perceived quality of video to be distributed over modern content distribution platforms, so that specific actions could be undertaken to maximise the Quality of Experience (QoE) of the users. Traditional VQA techniques based on feature extraction and modelling may not be sufficiently accurate. In this paper, a novel Deep Learning (DL) framework is introduced for effectively predicting VQA of video content delivery mechanisms based on end-to-end feature learning. The proposed framework is based on Convolutional Neural Networks, taking into account compression distortion as well as transmission delays. Training and evaluation of the proposed framework are performed on a user annotated VQA dataset specifically created to undertake this work. The experiments show that the proposed methods can lead to high accuracy of the quality estimation, showcasing the potential of using DL in complex VQA scenarios.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2018-09
issn 2331-8422
language eng
recordid cdi_proquest_journals_2112962371
source Publicly Available Content Database (Proquest) (PQ_SDU_P3)
subjects Artificial neural networks
Coding standards
Compression tests
Distortion
Feature extraction
Machine learning
Mathematical models
Neural networks
Quality
Quality assessment
User satisfaction
Video compression
Video transmission
title Convolutional Neural Networks for Video Quality Assessment
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T04%3A06%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Convolutional%20Neural%20Networks%20for%20Video%20Quality%20Assessment&rft.jtitle=arXiv.org&rft.au=Giannopoulos,%20Michalis&rft.date=2018-09-26&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2112962371%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_21129623713%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2112962371&rft_id=info:pmid/&rfr_iscdi=true