Loading…
The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems
Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deploy...
Saved in:
Published in: | IEEE eTransactions on network and service management 2022-12, Vol.19 (4), p.5152-5169 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23 |
---|---|
cites | cdi_FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23 |
container_end_page | 5169 |
container_issue | 4 |
container_start_page | 5152 |
container_title | IEEE eTransactions on network and service management |
container_volume | 19 |
creator | Apruzzese, Giovanni Pajola, Luca Conti, Mauro |
description | Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deployments, as well as on papers always relying on the same outdated data. The situation improved recently, as some efforts disclosed their labeled datasets. However, most past works used such datasets just as a 'yet another' testbed, overlooking the added potential provided by such availability. In contrast, we promote using such existing labeled data to cross-evaluate ML-NIDS. Such approach received only limited attention and, due to its complexity, requires a dedicated treatment. We hence propose the first cross-evaluation model. Our model highlights the broader range of realistic use-cases that can be assessed via cross-evaluations, allowing the discovery of still unknown qualities of state-of-the-art ML-NIDS. For instance, their detection surface can be extended-at no additional labeling cost. However, conducting such cross-evaluations is challenging. Hence, we propose the first framework, XeNIDS, for reliable cross-evaluations based on Network Flows. By using XeNIDS on six well-known datasets, we demonstrate the concealed potential, but also the risks, of cross-evaluations of ML-NIDS. |
doi_str_mv | 10.1109/TNSM.2022.3157344 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TNSM_2022_3157344</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9729769</ieee_id><sourcerecordid>2771521654</sourcerecordid><originalsourceid>FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23</originalsourceid><addsrcrecordid>eNpNkF1PwjAUhhujiYj-AOPNEq-H_VzXS0VUEsBE8LrpujMZwoZth-HfuwkxXjSnF897Ph6ErgkeEILV3WI2nw4opnTAiJCM8xPUI4rRmAsmT__9z9GF9yuMRUoU7aG3xRKioau9j0c7s25MKOsqqotoauyyrCCagHFVWX3ED8ZDHs0gfNfuMxpXwTW-Yx8hgP1Nzfc-wMZforPCrD1cHWsfvT-NFsOXePL6PB7eT2JLFQuxynJGTcqYLWSegrGWCZtkAmdYJJil7a4UaMITnmYYFwZYlmCccpZnWfso66PbQ9-tq78a8EGv6sZV7UhNpSSCkkTwliIHynZHOij01pUb4_aaYN2p05063anTR3Vt5uaQKQHgj1eSKpko9gMGcmor</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2771521654</pqid></control><display><type>article</type><title>The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Apruzzese, Giovanni ; Pajola, Luca ; Conti, Mauro</creator><creatorcontrib>Apruzzese, Giovanni ; Pajola, Luca ; Conti, Mauro</creatorcontrib><description>Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deployments, as well as on papers always relying on the same outdated data. The situation improved recently, as some efforts disclosed their labeled datasets. However, most past works used such datasets just as a 'yet another' testbed, overlooking the added potential provided by such availability. In contrast, we promote using such existing labeled data to cross-evaluate ML-NIDS. Such approach received only limited attention and, due to its complexity, requires a dedicated treatment. We hence propose the first cross-evaluation model. Our model highlights the broader range of realistic use-cases that can be assessed via cross-evaluations, allowing the discovery of still unknown qualities of state-of-the-art ML-NIDS. For instance, their detection surface can be extended-at no additional labeling cost. However, conducting such cross-evaluations is challenging. Hence, we propose the first framework, XeNIDS, for reliable cross-evaluations based on Network Flows. By using XeNIDS on six well-known datasets, we demonstrate the concealed potential, but also the risks, of cross-evaluations of ML-NIDS.</description><identifier>ISSN: 1932-4537</identifier><identifier>EISSN: 1932-4537</identifier><identifier>DOI: 10.1109/TNSM.2022.3157344</identifier><identifier>CODEN: ITNSC4</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Datasets ; Evaluation ; Intrusion detection systems ; Labeling ; Labels ; Machine learning ; Monitoring ; Network intrusion detection ; network security ; Proposals ; Reliability ; Supervised learning ; Training</subject><ispartof>IEEE eTransactions on network and service management, 2022-12, Vol.19 (4), p.5152-5169</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23</citedby><cites>FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23</cites><orcidid>0000-0002-3612-1934 ; 0000-0002-6890-9611</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9729769$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Apruzzese, Giovanni</creatorcontrib><creatorcontrib>Pajola, Luca</creatorcontrib><creatorcontrib>Conti, Mauro</creatorcontrib><title>The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems</title><title>IEEE eTransactions on network and service management</title><addtitle>T-NSM</addtitle><description>Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deployments, as well as on papers always relying on the same outdated data. The situation improved recently, as some efforts disclosed their labeled datasets. However, most past works used such datasets just as a 'yet another' testbed, overlooking the added potential provided by such availability. In contrast, we promote using such existing labeled data to cross-evaluate ML-NIDS. Such approach received only limited attention and, due to its complexity, requires a dedicated treatment. We hence propose the first cross-evaluation model. Our model highlights the broader range of realistic use-cases that can be assessed via cross-evaluations, allowing the discovery of still unknown qualities of state-of-the-art ML-NIDS. For instance, their detection surface can be extended-at no additional labeling cost. However, conducting such cross-evaluations is challenging. Hence, we propose the first framework, XeNIDS, for reliable cross-evaluations based on Network Flows. By using XeNIDS on six well-known datasets, we demonstrate the concealed potential, but also the risks, of cross-evaluations of ML-NIDS.</description><subject>Datasets</subject><subject>Evaluation</subject><subject>Intrusion detection systems</subject><subject>Labeling</subject><subject>Labels</subject><subject>Machine learning</subject><subject>Monitoring</subject><subject>Network intrusion detection</subject><subject>network security</subject><subject>Proposals</subject><subject>Reliability</subject><subject>Supervised learning</subject><subject>Training</subject><issn>1932-4537</issn><issn>1932-4537</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNpNkF1PwjAUhhujiYj-AOPNEq-H_VzXS0VUEsBE8LrpujMZwoZth-HfuwkxXjSnF897Ph6ErgkeEILV3WI2nw4opnTAiJCM8xPUI4rRmAsmT__9z9GF9yuMRUoU7aG3xRKioau9j0c7s25MKOsqqotoauyyrCCagHFVWX3ED8ZDHs0gfNfuMxpXwTW-Yx8hgP1Nzfc-wMZforPCrD1cHWsfvT-NFsOXePL6PB7eT2JLFQuxynJGTcqYLWSegrGWCZtkAmdYJJil7a4UaMITnmYYFwZYlmCccpZnWfso66PbQ9-tq78a8EGv6sZV7UhNpSSCkkTwliIHynZHOij01pUb4_aaYN2p05063anTR3Vt5uaQKQHgj1eSKpko9gMGcmor</recordid><startdate>202212</startdate><enddate>202212</enddate><creator>Apruzzese, Giovanni</creator><creator>Pajola, Luca</creator><creator>Conti, Mauro</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-3612-1934</orcidid><orcidid>https://orcid.org/0000-0002-6890-9611</orcidid></search><sort><creationdate>202212</creationdate><title>The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems</title><author>Apruzzese, Giovanni ; Pajola, Luca ; Conti, Mauro</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Datasets</topic><topic>Evaluation</topic><topic>Intrusion detection systems</topic><topic>Labeling</topic><topic>Labels</topic><topic>Machine learning</topic><topic>Monitoring</topic><topic>Network intrusion detection</topic><topic>network security</topic><topic>Proposals</topic><topic>Reliability</topic><topic>Supervised learning</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Apruzzese, Giovanni</creatorcontrib><creatorcontrib>Pajola, Luca</creatorcontrib><creatorcontrib>Conti, Mauro</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><jtitle>IEEE eTransactions on network and service management</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Apruzzese, Giovanni</au><au>Pajola, Luca</au><au>Conti, Mauro</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems</atitle><jtitle>IEEE eTransactions on network and service management</jtitle><stitle>T-NSM</stitle><date>2022-12</date><risdate>2022</risdate><volume>19</volume><issue>4</issue><spage>5152</spage><epage>5169</epage><pages>5152-5169</pages><issn>1932-4537</issn><eissn>1932-4537</eissn><coden>ITNSC4</coden><abstract>Enhancing Network Intrusion Detection Systems (NIDS) with supervised Machine Learning (ML) is tough. ML-NIDS must be trained and evaluated, operations requiring data where benign and malicious samples are clearly labeled. Such labels demand costly expert knowledge, resulting in a lack of real deployments, as well as on papers always relying on the same outdated data. The situation improved recently, as some efforts disclosed their labeled datasets. However, most past works used such datasets just as a 'yet another' testbed, overlooking the added potential provided by such availability. In contrast, we promote using such existing labeled data to cross-evaluate ML-NIDS. Such approach received only limited attention and, due to its complexity, requires a dedicated treatment. We hence propose the first cross-evaluation model. Our model highlights the broader range of realistic use-cases that can be assessed via cross-evaluations, allowing the discovery of still unknown qualities of state-of-the-art ML-NIDS. For instance, their detection surface can be extended-at no additional labeling cost. However, conducting such cross-evaluations is challenging. Hence, we propose the first framework, XeNIDS, for reliable cross-evaluations based on Network Flows. By using XeNIDS on six well-known datasets, we demonstrate the concealed potential, but also the risks, of cross-evaluations of ML-NIDS.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TNSM.2022.3157344</doi><tpages>18</tpages><orcidid>https://orcid.org/0000-0002-3612-1934</orcidid><orcidid>https://orcid.org/0000-0002-6890-9611</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1932-4537 |
ispartof | IEEE eTransactions on network and service management, 2022-12, Vol.19 (4), p.5152-5169 |
issn | 1932-4537 1932-4537 |
language | eng |
recordid | cdi_crossref_primary_10_1109_TNSM_2022_3157344 |
source | IEEE Electronic Library (IEL) Journals |
subjects | Datasets Evaluation Intrusion detection systems Labeling Labels Machine learning Monitoring Network intrusion detection network security Proposals Reliability Supervised learning Training |
title | The Cross-Evaluation of Machine Learning-Based Network Intrusion Detection Systems |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T18%3A37%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=The%20Cross-Evaluation%20of%20Machine%20Learning-Based%20Network%20Intrusion%20Detection%20Systems&rft.jtitle=IEEE%20eTransactions%20on%20network%20and%20service%20management&rft.au=Apruzzese,%20Giovanni&rft.date=2022-12&rft.volume=19&rft.issue=4&rft.spage=5152&rft.epage=5169&rft.pages=5152-5169&rft.issn=1932-4537&rft.eissn=1932-4537&rft.coden=ITNSC4&rft_id=info:doi/10.1109/TNSM.2022.3157344&rft_dat=%3Cproquest_cross%3E2771521654%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c293t-9bd32a833cf7d8eacc35c6b50b0560385372e264648b00fae3b600843dbb3db23%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2771521654&rft_id=info:pmid/&rft_ieee_id=9729769&rfr_iscdi=true |