Loading…

An overview and a benchmark of active learning for outlier detection with one-class classifiers

Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2021-04, Vol.168, p.114372, Article 114372
Main Authors: Trittenbach, Holger, Englhardt, Adrian, Böhm, Klemens
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283
cites cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283
container_end_page
container_issue
container_start_page 114372
container_title Expert systems with applications
container_volume 168
creator Trittenbach, Holger
Englhardt, Adrian
Böhm, Klemens
description Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article. This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers. •Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.
doi_str_mv 10.1016/j.eswa.2020.114372
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2492316448</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417420310496</els_id><sourcerecordid>2492316448</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</originalsourceid><addsrcrecordid>eNp9kE9PAyEQxYnRxFr9Ap5IPG8FloXdxEvT-C9p4kXPhMJgWdelwraN317W9exlSIb33sz8ELqmZEEJFbftAtJRLxhhuUF5KdkJmtFaloWQTXmKZqSpZMGp5OfoIqWWECoJkTOklj0OB4gHD0ese4s13kBvtp86fuDgsDaDPwDuQMfe9-_YhYjDfug8RGxhgPwdenz0wxaHHgrT6ZTwb_Uua9IlOnO6S3D1987R28P96-qpWL88Pq-W68LkXYeirLmQ1HBhNW-YM7WUXAjLbSWFMSUVDasIpSUFJzYVWK4dNFrWdSMqAqwu5-hmyt3F8LWHNKg27GOfRyqWE3MC56OKTSoTQ0oRnNpFn0_9VpSoEaRq1QhSjSDVBDKb7iYT5P0zp6iS8ZkRWB_z_coG_5_9B4UZe7k</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2492316448</pqid></control><display><type>article</type><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</creator><creatorcontrib>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</creatorcontrib><description>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article. This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers. •Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2020.114372</identifier><language>eng</language><publisher>New York: Elsevier Ltd</publisher><subject>Active learning ; Anomaly detection ; Business competition ; Classification ; Classifiers ; Data analysis ; Evaluation ; Feedback ; Learning ; One-class classification ; Outlier detection ; Outliers (statistics) ; Teaching methods ; User feedback</subject><ispartof>Expert systems with applications, 2021-04, Vol.168, p.114372, Article 114372</ispartof><rights>2020 Elsevier Ltd</rights><rights>Copyright Elsevier BV Apr 15, 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</citedby><cites>FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Trittenbach, Holger</creatorcontrib><creatorcontrib>Englhardt, Adrian</creatorcontrib><creatorcontrib>Böhm, Klemens</creatorcontrib><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><title>Expert systems with applications</title><description>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article. This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers. •Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</description><subject>Active learning</subject><subject>Anomaly detection</subject><subject>Business competition</subject><subject>Classification</subject><subject>Classifiers</subject><subject>Data analysis</subject><subject>Evaluation</subject><subject>Feedback</subject><subject>Learning</subject><subject>One-class classification</subject><subject>Outlier detection</subject><subject>Outliers (statistics)</subject><subject>Teaching methods</subject><subject>User feedback</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kE9PAyEQxYnRxFr9Ap5IPG8FloXdxEvT-C9p4kXPhMJgWdelwraN317W9exlSIb33sz8ELqmZEEJFbftAtJRLxhhuUF5KdkJmtFaloWQTXmKZqSpZMGp5OfoIqWWECoJkTOklj0OB4gHD0ese4s13kBvtp86fuDgsDaDPwDuQMfe9-_YhYjDfug8RGxhgPwdenz0wxaHHgrT6ZTwb_Uua9IlOnO6S3D1987R28P96-qpWL88Pq-W68LkXYeirLmQ1HBhNW-YM7WUXAjLbSWFMSUVDasIpSUFJzYVWK4dNFrWdSMqAqwu5-hmyt3F8LWHNKg27GOfRyqWE3MC56OKTSoTQ0oRnNpFn0_9VpSoEaRq1QhSjSDVBDKb7iYT5P0zp6iS8ZkRWB_z_coG_5_9B4UZe7k</recordid><startdate>20210415</startdate><enddate>20210415</enddate><creator>Trittenbach, Holger</creator><creator>Englhardt, Adrian</creator><creator>Böhm, Klemens</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20210415</creationdate><title>An overview and a benchmark of active learning for outlier detection with one-class classifiers</title><author>Trittenbach, Holger ; Englhardt, Adrian ; Böhm, Klemens</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Active learning</topic><topic>Anomaly detection</topic><topic>Business competition</topic><topic>Classification</topic><topic>Classifiers</topic><topic>Data analysis</topic><topic>Evaluation</topic><topic>Feedback</topic><topic>Learning</topic><topic>One-class classification</topic><topic>Outlier detection</topic><topic>Outliers (statistics)</topic><topic>Teaching methods</topic><topic>User feedback</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Trittenbach, Holger</creatorcontrib><creatorcontrib>Englhardt, Adrian</creatorcontrib><creatorcontrib>Böhm, Klemens</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Trittenbach, Holger</au><au>Englhardt, Adrian</au><au>Böhm, Klemens</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An overview and a benchmark of active learning for outlier detection with one-class classifiers</atitle><jtitle>Expert systems with applications</jtitle><date>2021-04-15</date><risdate>2021</risdate><volume>168</volume><spage>114372</spage><pages>114372-</pages><artnum>114372</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>Active learning methods increase classification quality by means of user feedback. An important subcategory is active learning for outlier detection with one-class classifiers. While various methods in this category exist, selecting one for a given application scenario is difficult. This is because existing methods rely on different assumptions, have different objectives, and often are tailored to a specific use case. All this calls for a comprehensive comparison, the topic of this article. This article starts with a categorization of the various methods. Interestingly, many assumptions in the literature are implicit, and their impact has not been discussed so far. Based on this, we propose a novel approach to evaluate active learning results by quantifying how classification results evolve with more user feedback, in a compact and nuanced manner. We run over 84,000 experiments to compare state-of-the-art one-class active learning methods, for a broad variety of scenarios. One key finding is that there is no single active learning method that stands out in a competitive evaluation. Instead, we found that selecting a good query strategy alone is not sufficient, since results hinge significantly on other factors, such as the selection of hyperparameter values. Our results show that some configurations are more robust than others. We conclude by phrasing our findings as guidelines on how to select active learning methods for outlier detection with one-class classifiers. •Categorization of assumptions and objectives of one-class active learning.•Novel progress curve summaries to facilitate reliable evaluation of active learning.•Large benchmark with 84,000 learning scenarios, classifiers, and query strategies.•Derivation of guidelines to select suitable one-class active learning methods.</abstract><cop>New York</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2020.114372</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2021-04, Vol.168, p.114372, Article 114372
issn 0957-4174
1873-6793
language eng
recordid cdi_proquest_journals_2492316448
source ScienceDirect Freedom Collection 2022-2024
subjects Active learning
Anomaly detection
Business competition
Classification
Classifiers
Data analysis
Evaluation
Feedback
Learning
One-class classification
Outlier detection
Outliers (statistics)
Teaching methods
User feedback
title An overview and a benchmark of active learning for outlier detection with one-class classifiers
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-05T00%3A21%3A09IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20overview%20and%20a%20benchmark%20of%20active%20learning%20for%20outlier%20detection%20with%20one-class%20classifiers&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Trittenbach,%20Holger&rft.date=2021-04-15&rft.volume=168&rft.spage=114372&rft.pages=114372-&rft.artnum=114372&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2020.114372&rft_dat=%3Cproquest_cross%3E2492316448%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c372t-384671c46da492fc877466d4d576cc31692501131ef6b5ed4afe9a7889650e283%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2492316448&rft_id=info:pmid/&rfr_iscdi=true