Loading…
An overview and a benchmark of active learning for outlier detection with one-class classifiers
Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because...
Saved in:
Published in: | Expert systems with applications 2021-04, Vol.168, p.114372, Article 114372 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283 |
---|---|
cites | cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283 |
container_end_page | |
container_issue | |
container_start_page | 114372 |
container_title | Expert systems with applications |
container_volume | 168 |
creator | Trittenbach, Holger Englhardt, Adrian Böhm, Klemens |
description | Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article.
This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers.
•Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods. |
doi_str_mv | 10.1016/j.eswa.2020.114372 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2492316448</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417420310496</els_id><sourcerecordid>2492316448</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</originalsourceid><addsrcrecordid>eNp9kE9PAyEQxYnRxFr9Ap5IPG8FloXdxEvT-C9p4kXPhMJgWdelwraN317W9exlSIb33sz8ELqmZEEJFbftAtJRLxhhuUF5KdkJmtFaloWQTXmKZqSpZMGp5OfoIqWWECoJkTOklj0OB4gHD0ese4s13kBvtp86fuDgsDaDPwDuQMfe9-_YhYjDfug8RGxhgPwdenz0wxaHHgrT6ZTwb_Uua9IlOnO6S3D1987R28P96-qpWL88Pq-W68LkXYeirLmQ1HBhNW-YM7WUXAjLbSWFMSUVDasIpSUFJzYVWK4dNFrWdSMqAqwu5-hmyt3F8LWHNKg27GOfRyqWE3MC56OKTSoTQ0oRnNpFn0_9VpSoEaRq1QhSjSDVBDKb7iYT5P0zp6iS8ZkRWB_z_coG_5_9B4UZe7k</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2492316448</pqid></control><display><type>article</type><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</creator><creatorcontrib>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</creatorcontrib><description>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article.
This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers.
•Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2020.114372</identifier><language>eng</language><publisher>New York: Elsevier Ltd</publisher><subject>Active learning ; Anomaly detection ; Business competition ; Classification ; Classifiers ; Data analysis ; Evaluation ; Feedback ; Learning ; One-class classification ; Outlier detection ; Outliers (statistics) ; Teaching methods ; User feedback</subject><ispartof>Expert systems with applications, 2021-04, Vol.168, p.114372, Article 114372</ispartof><rights>2020 Elsevier Ltd</rights><rights>Copyright Elsevier BV Apr 15, 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</citedby><cites>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Trittenbach, Holger</creatorcontrib><creatorcontrib>Englhardt, Adrian</creatorcontrib><creatorcontrib>Böhm, Klemens</creatorcontrib><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><title>Expert systems with applications</title><description>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article.
This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers.
•Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</description><subject>Active learning</subject><subject>Anomaly detection</subject><subject>Business competition</subject><subject>Classification</subject><subject>Classifiers</subject><subject>Data analysis</subject><subject>Evaluation</subject><subject>Feedback</subject><subject>Learning</subject><subject>One-class classification</subject><subject>Outlier detection</subject><subject>Outliers (statistics)</subject><subject>Teaching methods</subject><subject>User feedback</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kE9PAyEQxYnRxFr9Ap5IPG8FloXdxEvT-C9p4kXPhMJgWdelwraN317W9exlSIb33sz8ELqmZEEJFbftAtJRLxhhuUF5KdkJmtFaloWQTXmKZqSpZMGp5OfoIqWWECoJkTOklj0OB4gHD0ese4s13kBvtp86fuDgsDaDPwDuQMfe9-_YhYjDfug8RGxhgPwdenz0wxaHHgrT6ZTwb_Uua9IlOnO6S3D1987R28P96-qpWL88Pq-W68LkXYeirLmQ1HBhNW-YM7WUXAjLbSWFMSUVDasIpSUFJzYVWK4dNFrWdSMqAqwu5-hmyt3F8LWHNKg27GOfRyqWE3MC56OKTSoTQ0oRnNpFn0_9VpSoEaRq1QhSjSDVBDKb7iYT5P0zp6iS8ZkRWB_z_coG_5_9B4UZe7k</recordid><startdate>20210415</startdate><enddate>20210415</enddate><creator>Trittenbach, Holger</creator><creator>Englhardt, Adrian</creator><creator>Böhm, Klemens</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20210415</creationdate><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><author>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Active learning</topic><topic>Anomaly detection</topic><topic>Business competition</topic><topic>Classification</topic><topic>Classifiers</topic><topic>Data analysis</topic><topic>Evaluation</topic><topic>Feedback</topic><topic>Learning</topic><topic>One-class classification</topic><topic>Outlier detection</topic><topic>Outliers (statistics)</topic><topic>Teaching methods</topic><topic>User feedback</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Trittenbach, Holger</creatorcontrib><creatorcontrib>Englhardt, Adrian</creatorcontrib><creatorcontrib>Böhm, Klemens</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Trittenbach, Holger</au><au>Englhardt, Adrian</au><au>Böhm, Klemens</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An overview and a benchmark of active learning for outlier detection with one-class classifiers</atitle><jtitle>Expert systems with applications</jtitle><date>2021-04-15</date><risdate>2021</risdate><volume>168</volume><spage>114372</spage><pages>114372-</pages><artnum>114372</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article.
This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers.
•Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</abstract><cop>New York</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2020.114372</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0957-4174 |
ispartof | Expert systems with applications, 2021-04, Vol.168, p.114372, Article 114372 |
issn | 0957-4174 1873-6793 |
language | eng |
recordid | cdi_proquest_journals_2492316448 |
source | ScienceDirect Freedom Collection 2022-2024 |
subjects | Active learning Anomaly detection Business competition Classification Classifiers Data analysis Evaluation Feedback Learning One-class classification Outlier detection Outliers (statistics) Teaching methods User feedback |
title | An overview and a benchmark of active learning for outlier detection with one-class classifiers |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T00%3A21%3A09IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20overview%20and%20a%20benchmark%20of%20active%20learning%20for%20outlier%20detection%20with%20one-class%20classifiers&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Trittenbach,%20Holger&rft.date=2021-04-15&rft.volume=168&rft.spage=114372&rft.pages=114372-&rft.artnum=114372&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2020.114372&rft_dat=%3Cproquest_cross%3E2492316448%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2492316448&rft_id=info:pmid/&rfr_iscdi=true |