Loading…

Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures

Modern networks, characterized by their complexity and heterogeneity, are transitioning from manual to automated and intelligent management, according to the vision of Autonomous Networks (ANs). Leveraging data-driven techniques, ANs aim to provide the "Zero-X" and "Self-X" exper...

Full description

Saved in:
Bibliographic Details
Main Authors: Ave, Thomas, Soto, Paola, Camelo, Miguel, De Schepper, Tom, Mets, Kevin
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 7
container_issue
container_start_page 1
container_title
container_volume
creator Ave, Thomas
Soto, Paola
Camelo, Miguel
De Schepper, Tom
Mets, Kevin
description Modern networks, characterized by their complexity and heterogeneity, are transitioning from manual to automated and intelligent management, according to the vision of Autonomous Networks (ANs). Leveraging data-driven techniques, ANs aim to provide the "Zero-X" and "Self-X" experience, where intelligent and adaptable network operations are key cornerstones. Such is the case of intelligent resource scaling, where the goal is to optimize resource orchestration to maximize efficiency, reduce latency, and maintain high-quality service, even amid fluctuating network loads and changing service requirements. Unfortunately, current approaches for auto-scaling are computationally expensive to deploy in resource-constrained devices such as those found at the edge or beyond. This paper introduces an innovative four-phase approach to train a compact, data-driven Deep Reinforcement Learning (DRL) scaler that can be deployed on low-power devices. Our results demonstrate the scalability and efficiency of this model, achieving state-of-the-art scaling with up to 1003x fewer parameters, enhancing interpretability and computational efficiency, making it a robust solution for intelligent resource scaling in network environments. The resulting 1487x runtime speed improvement and 28.5x reduction in memory requirements allow the scaler to be deployed on low-power devices and still operate in real-time, which is essential for mission-critical and latency-sensitive applications.
doi_str_mv 10.1109/NOMS59830.2024.10575377
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_10575377</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10575377</ieee_id><sourcerecordid>10575377</sourcerecordid><originalsourceid>FETCH-LOGICAL-i134t-b4cdb927d4c9cc11d5f250b475dd20b82167b3352042128e4394353cccf7a5b53</originalsourceid><addsrcrecordid>eNo1kNFKwzAUQKMgOOf-QDA_0HmTm5jlcQ6dg7kNqm_CaNPbGe2akUbK_t6B-nTgPJyHw9itgLEQYO9W65dc2wnCWIJUYwHaaDTmjI2sOWkNKI1Fe84GEo3KrAF7ya667hNAGUAYsPdNaLw78lnYHyJ1nQ8tr0Pky9Bnm9BT5Is2UdP4HbWJ565ofLvjvuV5qFNfRMoeio4qvqLUh_jFp9F9-EQufZ9q1-yiLpqORn8csrenx9fZc7Zczxez6TLzAlXKSuWq0kpTKWedE6LStdRQKqOrSkI5keLelIhagpJCTkihVajROVebQpcah-zmt-uJaHuIfl_E4_b_Bv4A8xNVMg</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures</title><source>IEEE Xplore All Conference Series</source><creator>Ave, Thomas ; Soto, Paola ; Camelo, Miguel ; De Schepper, Tom ; Mets, Kevin</creator><creatorcontrib>Ave, Thomas ; Soto, Paola ; Camelo, Miguel ; De Schepper, Tom ; Mets, Kevin</creatorcontrib><description>Modern networks, characterized by their complexity and heterogeneity, are transitioning from manual to automated and intelligent management, according to the vision of Autonomous Networks (ANs). Leveraging data-driven techniques, ANs aim to provide the "Zero-X" and "Self-X" experience, where intelligent and adaptable network operations are key cornerstones. Such is the case of intelligent resource scaling, where the goal is to optimize resource orchestration to maximize efficiency, reduce latency, and maintain high-quality service, even amid fluctuating network loads and changing service requirements. Unfortunately, current approaches for auto-scaling are computationally expensive to deploy in resource-constrained devices such as those found at the edge or beyond. This paper introduces an innovative four-phase approach to train a compact, data-driven Deep Reinforcement Learning (DRL) scaler that can be deployed on low-power devices. Our results demonstrate the scalability and efficiency of this model, achieving state-of-the-art scaling with up to 1003x fewer parameters, enhancing interpretability and computational efficiency, making it a robust solution for intelligent resource scaling in network environments. The resulting 1487x runtime speed improvement and 28.5x reduction in memory requirements allow the scaler to be deployed on low-power devices and still operate in real-time, which is essential for mission-critical and latency-sensitive applications.</description><identifier>EISSN: 2374-9709</identifier><identifier>EISBN: 9798350327939</identifier><identifier>DOI: 10.1109/NOMS59830.2024.10575377</identifier><language>eng</language><publisher>IEEE</publisher><subject>Adaptation models ; Computational modeling ; Memory management ; Network architecture ; Runtime ; Scalability ; Training</subject><ispartof>NOMS 2024-2024 IEEE Network Operations and Management Symposium, 2024, p.1-7</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10575377$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,27904,54533,54910</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10575377$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Ave, Thomas</creatorcontrib><creatorcontrib>Soto, Paola</creatorcontrib><creatorcontrib>Camelo, Miguel</creatorcontrib><creatorcontrib>De Schepper, Tom</creatorcontrib><creatorcontrib>Mets, Kevin</creatorcontrib><title>Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures</title><title>NOMS 2024-2024 IEEE Network Operations and Management Symposium</title><addtitle>NOMS</addtitle><description>Modern networks, characterized by their complexity and heterogeneity, are transitioning from manual to automated and intelligent management, according to the vision of Autonomous Networks (ANs). Leveraging data-driven techniques, ANs aim to provide the "Zero-X" and "Self-X" experience, where intelligent and adaptable network operations are key cornerstones. Such is the case of intelligent resource scaling, where the goal is to optimize resource orchestration to maximize efficiency, reduce latency, and maintain high-quality service, even amid fluctuating network loads and changing service requirements. Unfortunately, current approaches for auto-scaling are computationally expensive to deploy in resource-constrained devices such as those found at the edge or beyond. This paper introduces an innovative four-phase approach to train a compact, data-driven Deep Reinforcement Learning (DRL) scaler that can be deployed on low-power devices. Our results demonstrate the scalability and efficiency of this model, achieving state-of-the-art scaling with up to 1003x fewer parameters, enhancing interpretability and computational efficiency, making it a robust solution for intelligent resource scaling in network environments. The resulting 1487x runtime speed improvement and 28.5x reduction in memory requirements allow the scaler to be deployed on low-power devices and still operate in real-time, which is essential for mission-critical and latency-sensitive applications.</description><subject>Adaptation models</subject><subject>Computational modeling</subject><subject>Memory management</subject><subject>Network architecture</subject><subject>Runtime</subject><subject>Scalability</subject><subject>Training</subject><issn>2374-9709</issn><isbn>9798350327939</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2024</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNo1kNFKwzAUQKMgOOf-QDA_0HmTm5jlcQ6dg7kNqm_CaNPbGe2akUbK_t6B-nTgPJyHw9itgLEQYO9W65dc2wnCWIJUYwHaaDTmjI2sOWkNKI1Fe84GEo3KrAF7ya667hNAGUAYsPdNaLw78lnYHyJ1nQ8tr0Pky9Bnm9BT5Is2UdP4HbWJ565ofLvjvuV5qFNfRMoeio4qvqLUh_jFp9F9-EQufZ9q1-yiLpqORn8csrenx9fZc7Zczxez6TLzAlXKSuWq0kpTKWedE6LStdRQKqOrSkI5keLelIhagpJCTkihVajROVebQpcah-zmt-uJaHuIfl_E4_b_Bv4A8xNVMg</recordid><startdate>20240506</startdate><enddate>20240506</enddate><creator>Ave, Thomas</creator><creator>Soto, Paola</creator><creator>Camelo, Miguel</creator><creator>De Schepper, Tom</creator><creator>Mets, Kevin</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>20240506</creationdate><title>Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures</title><author>Ave, Thomas ; Soto, Paola ; Camelo, Miguel ; De Schepper, Tom ; Mets, Kevin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i134t-b4cdb927d4c9cc11d5f250b475dd20b82167b3352042128e4394353cccf7a5b53</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Adaptation models</topic><topic>Computational modeling</topic><topic>Memory management</topic><topic>Network architecture</topic><topic>Runtime</topic><topic>Scalability</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Ave, Thomas</creatorcontrib><creatorcontrib>Soto, Paola</creatorcontrib><creatorcontrib>Camelo, Miguel</creatorcontrib><creatorcontrib>De Schepper, Tom</creatorcontrib><creatorcontrib>Mets, Kevin</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Ave, Thomas</au><au>Soto, Paola</au><au>Camelo, Miguel</au><au>De Schepper, Tom</au><au>Mets, Kevin</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures</atitle><btitle>NOMS 2024-2024 IEEE Network Operations and Management Symposium</btitle><stitle>NOMS</stitle><date>2024-05-06</date><risdate>2024</risdate><spage>1</spage><epage>7</epage><pages>1-7</pages><eissn>2374-9709</eissn><eisbn>9798350327939</eisbn><abstract>Modern networks, characterized by their complexity and heterogeneity, are transitioning from manual to automated and intelligent management, according to the vision of Autonomous Networks (ANs). Leveraging data-driven techniques, ANs aim to provide the "Zero-X" and "Self-X" experience, where intelligent and adaptable network operations are key cornerstones. Such is the case of intelligent resource scaling, where the goal is to optimize resource orchestration to maximize efficiency, reduce latency, and maintain high-quality service, even amid fluctuating network loads and changing service requirements. Unfortunately, current approaches for auto-scaling are computationally expensive to deploy in resource-constrained devices such as those found at the edge or beyond. This paper introduces an innovative four-phase approach to train a compact, data-driven Deep Reinforcement Learning (DRL) scaler that can be deployed on low-power devices. Our results demonstrate the scalability and efficiency of this model, achieving state-of-the-art scaling with up to 1003x fewer parameters, enhancing interpretability and computational efficiency, making it a robust solution for intelligent resource scaling in network environments. The resulting 1487x runtime speed improvement and 28.5x reduction in memory requirements allow the scaler to be deployed on low-power devices and still operate in real-time, which is essential for mission-critical and latency-sensitive applications.</abstract><pub>IEEE</pub><doi>10.1109/NOMS59830.2024.10575377</doi><tpages>7</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2374-9709
ispartof NOMS 2024-2024 IEEE Network Operations and Management Symposium, 2024, p.1-7
issn 2374-9709
language eng
recordid cdi_ieee_primary_10575377
source IEEE Xplore All Conference Series
subjects Adaptation models
Computational modeling
Memory management
Network architecture
Runtime
Scalability
Training
title Policy Compression for Low-Power Intelligent Scaling in Software-Based Network Architectures
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-25T23%3A12%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Policy%20Compression%20for%20Low-Power%20Intelligent%20Scaling%20in%20Software-Based%20Network%20Architectures&rft.btitle=NOMS%202024-2024%20IEEE%20Network%20Operations%20and%20Management%20Symposium&rft.au=Ave,%20Thomas&rft.date=2024-05-06&rft.spage=1&rft.epage=7&rft.pages=1-7&rft.eissn=2374-9709&rft_id=info:doi/10.1109/NOMS59830.2024.10575377&rft.eisbn=9798350327939&rft_dat=%3Cieee_CHZPO%3E10575377%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i134t-b4cdb927d4c9cc11d5f250b475dd20b82167b3352042128e4394353cccf7a5b53%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10575377&rfr_iscdi=true