Loading…

FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics

Detection-driven real-time video analytics require continuous detection of objects contained in the video frames using deep learning models like YOLOV3, EfficientDet. However, running these detectors on each and every frame in resource-constrained edge devices is computationally intensive. By taking...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2022-03
Main Authors: Arefeen, Md Adnan, Sumaiya Tabassum Nimi, Md Yusuf Sarwar Uddin
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Arefeen, Md Adnan
Sumaiya Tabassum Nimi
Md Yusuf Sarwar Uddin
description Detection-driven real-time video analytics require continuous detection of objects contained in the video frames using deep learning models like YOLOV3, EfficientDet. However, running these detectors on each and every frame in resource-constrained edge devices is computationally intensive. By taking the temporal correlation between consecutive video frames into account, we note that detection outputs tend to be overlapping in successive frames. Elimination of similar consecutive frames will lead to a negligible drop in performance while offering significant performance benefits by reducing overall computation and communication costs. The key technical questions are, therefore, (a) how to identify which frames to be processed by the object detector, and (b) how many successive frames can be skipped (called skip-length) once a frame is selected to be processed. The overall goal of the process is to keep the error due to skipping frames as small as possible. We introduce a novel error vs processing rate optimization problem with respect to the object detection task that balances between the error rate and the fraction of frames filtering. Subsequently, we propose an off-line Reinforcement Learning (RL)-based algorithm to determine these skip-lengths as a state-action policy of the RL agent from a recorded video and then deploy the agent online for live video streams. To this end, we develop FrameHopper, an edge-cloud collaborative video analytics framework, that runs a lightweight trained RL agent on the camera and passes filtered frames to the server where the object detection model runs for a set of applications. We have tested our approach on a number of live videos captured from real-life scenarios and show that FrameHopper processes only a handful of frames but produces detection results closer to the oracle solution and outperforms recent state-of-the-art solutions in most cases.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2642083857</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2642083857</sourcerecordid><originalsourceid>FETCH-proquest_journals_26420838573</originalsourceid><addsrcrecordid>eNqNyk0LgjAYAOARBEn5HwadB7b5RbeoxGOUdBXR15jMzfbOoH-fhT-g03N5FsTjQuxYGnK-Ij5iFwQBjxMeRcIjZWarHnIzDGD39AYKaidfQC_W1IAo9YOalt5lA4b-KlKp6Qnc9xnNGjttTa9QKVbIHuZ60JV6O1njhizbSiH4s2uyzc7FMWeDNc8R0JWdGe20seRxyINUpFEi_lsfXS9EBg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2642083857</pqid></control><display><type>article</type><title>FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics</title><source>Publicly Available Content Database</source><creator>Arefeen, Md Adnan ; Sumaiya Tabassum Nimi ; Md Yusuf Sarwar Uddin</creator><creatorcontrib>Arefeen, Md Adnan ; Sumaiya Tabassum Nimi ; Md Yusuf Sarwar Uddin</creatorcontrib><description>Detection-driven real-time video analytics require continuous detection of objects contained in the video frames using deep learning models like YOLOV3, EfficientDet. However, running these detectors on each and every frame in resource-constrained edge devices is computationally intensive. By taking the temporal correlation between consecutive video frames into account, we note that detection outputs tend to be overlapping in successive frames. Elimination of similar consecutive frames will lead to a negligible drop in performance while offering significant performance benefits by reducing overall computation and communication costs. The key technical questions are, therefore, (a) how to identify which frames to be processed by the object detector, and (b) how many successive frames can be skipped (called skip-length) once a frame is selected to be processed. The overall goal of the process is to keep the error due to skipping frames as small as possible. We introduce a novel error vs processing rate optimization problem with respect to the object detection task that balances between the error rate and the fraction of frames filtering. Subsequently, we propose an off-line Reinforcement Learning (RL)-based algorithm to determine these skip-lengths as a state-action policy of the RL agent from a recorded video and then deploy the agent online for live video streams. To this end, we develop FrameHopper, an edge-cloud collaborative video analytics framework, that runs a lightweight trained RL agent on the camera and passes filtered frames to the server where the object detection model runs for a set of applications. We have tested our approach on a number of live videos captured from real-life scenarios and show that FrameHopper processes only a handful of frames but produces detection results closer to the oracle solution and outperforms recent state-of-the-art solutions in most cases.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Deep learning ; Frames (data processing) ; Machine learning ; Microbalances ; Object recognition ; Optimization ; Real time ; Video data</subject><ispartof>arXiv.org, 2022-03</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2642083857?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25751,37010,44588</link.rule.ids></links><search><creatorcontrib>Arefeen, Md Adnan</creatorcontrib><creatorcontrib>Sumaiya Tabassum Nimi</creatorcontrib><creatorcontrib>Md Yusuf Sarwar Uddin</creatorcontrib><title>FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics</title><title>arXiv.org</title><description>Detection-driven real-time video analytics require continuous detection of objects contained in the video frames using deep learning models like YOLOV3, EfficientDet. However, running these detectors on each and every frame in resource-constrained edge devices is computationally intensive. By taking the temporal correlation between consecutive video frames into account, we note that detection outputs tend to be overlapping in successive frames. Elimination of similar consecutive frames will lead to a negligible drop in performance while offering significant performance benefits by reducing overall computation and communication costs. The key technical questions are, therefore, (a) how to identify which frames to be processed by the object detector, and (b) how many successive frames can be skipped (called skip-length) once a frame is selected to be processed. The overall goal of the process is to keep the error due to skipping frames as small as possible. We introduce a novel error vs processing rate optimization problem with respect to the object detection task that balances between the error rate and the fraction of frames filtering. Subsequently, we propose an off-line Reinforcement Learning (RL)-based algorithm to determine these skip-lengths as a state-action policy of the RL agent from a recorded video and then deploy the agent online for live video streams. To this end, we develop FrameHopper, an edge-cloud collaborative video analytics framework, that runs a lightweight trained RL agent on the camera and passes filtered frames to the server where the object detection model runs for a set of applications. We have tested our approach on a number of live videos captured from real-life scenarios and show that FrameHopper processes only a handful of frames but produces detection results closer to the oracle solution and outperforms recent state-of-the-art solutions in most cases.</description><subject>Algorithms</subject><subject>Deep learning</subject><subject>Frames (data processing)</subject><subject>Machine learning</subject><subject>Microbalances</subject><subject>Object recognition</subject><subject>Optimization</subject><subject>Real time</subject><subject>Video data</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyk0LgjAYAOARBEn5HwadB7b5RbeoxGOUdBXR15jMzfbOoH-fhT-g03N5FsTjQuxYGnK-Ij5iFwQBjxMeRcIjZWarHnIzDGD39AYKaidfQC_W1IAo9YOalt5lA4b-KlKp6Qnc9xnNGjttTa9QKVbIHuZ60JV6O1njhizbSiH4s2uyzc7FMWeDNc8R0JWdGe20seRxyINUpFEi_lsfXS9EBg</recordid><startdate>20220322</startdate><enddate>20220322</enddate><creator>Arefeen, Md Adnan</creator><creator>Sumaiya Tabassum Nimi</creator><creator>Md Yusuf Sarwar Uddin</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20220322</creationdate><title>FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics</title><author>Arefeen, Md Adnan ; Sumaiya Tabassum Nimi ; Md Yusuf Sarwar Uddin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_26420838573</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Deep learning</topic><topic>Frames (data processing)</topic><topic>Machine learning</topic><topic>Microbalances</topic><topic>Object recognition</topic><topic>Optimization</topic><topic>Real time</topic><topic>Video data</topic><toplevel>online_resources</toplevel><creatorcontrib>Arefeen, Md Adnan</creatorcontrib><creatorcontrib>Sumaiya Tabassum Nimi</creatorcontrib><creatorcontrib>Md Yusuf Sarwar Uddin</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Arefeen, Md Adnan</au><au>Sumaiya Tabassum Nimi</au><au>Md Yusuf Sarwar Uddin</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics</atitle><jtitle>arXiv.org</jtitle><date>2022-03-22</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>Detection-driven real-time video analytics require continuous detection of objects contained in the video frames using deep learning models like YOLOV3, EfficientDet. However, running these detectors on each and every frame in resource-constrained edge devices is computationally intensive. By taking the temporal correlation between consecutive video frames into account, we note that detection outputs tend to be overlapping in successive frames. Elimination of similar consecutive frames will lead to a negligible drop in performance while offering significant performance benefits by reducing overall computation and communication costs. The key technical questions are, therefore, (a) how to identify which frames to be processed by the object detector, and (b) how many successive frames can be skipped (called skip-length) once a frame is selected to be processed. The overall goal of the process is to keep the error due to skipping frames as small as possible. We introduce a novel error vs processing rate optimization problem with respect to the object detection task that balances between the error rate and the fraction of frames filtering. Subsequently, we propose an off-line Reinforcement Learning (RL)-based algorithm to determine these skip-lengths as a state-action policy of the RL agent from a recorded video and then deploy the agent online for live video streams. To this end, we develop FrameHopper, an edge-cloud collaborative video analytics framework, that runs a lightweight trained RL agent on the camera and passes filtered frames to the server where the object detection model runs for a set of applications. We have tested our approach on a number of live videos captured from real-life scenarios and show that FrameHopper processes only a handful of frames but produces detection results closer to the oracle solution and outperforms recent state-of-the-art solutions in most cases.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-03
issn 2331-8422
language eng
recordid cdi_proquest_journals_2642083857
source Publicly Available Content Database
subjects Algorithms
Deep learning
Frames (data processing)
Machine learning
Microbalances
Object recognition
Optimization
Real time
Video data
title FrameHopper: Selective Processing of Video Frames in Detection-driven Real-Time Video Analytics
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T03%3A31%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=FrameHopper:%20Selective%20Processing%20of%20Video%20Frames%20in%20Detection-driven%20Real-Time%20Video%20Analytics&rft.jtitle=arXiv.org&rft.au=Arefeen,%20Md%20Adnan&rft.date=2022-03-22&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2642083857%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_26420838573%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2642083857&rft_id=info:pmid/&rfr_iscdi=true