Loading…
CIML-R: Causally Informed Machine Learning Based on Feature Relevance
Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domai...
Saved in:
Main Authors: | , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 75 |
container_issue | |
container_start_page | 68 |
container_title | |
container_volume | |
creator | Surner, Martin Khelil, Abdelmajid |
description | Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations. |
doi_str_mv | 10.1109/SDS60720.2024.00018 |
format | conference_proceeding |
fullrecord | <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_10675979</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10675979</ieee_id><sourcerecordid>10675979</sourcerecordid><originalsourceid>FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3</originalsourceid><addsrcrecordid>eNotT9tKw0AUXAXBUvMF-rA_kHj2ls36prHVQIrQ6nM52T3RSJpK0gr9-y7oywwzDMMMY7cCMiHA3W-eNzlYCZkEqTMAEMUFS5x1hTKgwEmnL9lMRpUqLeGaJdP0HWNKCmF1MWOLslrV6fqBl3icsO9PvBra_bijwFfov7qBeE04Dt3wyZ9wivZ-4EvCw3EkvqaefnHwdMOuWuwnSv55zj6Wi_fyNa3fXqrysU47AfkhNarxgnyjPHojEQ21KpAN0kdU2hYmKB-HFY0wOlivQmt9aIUTULgcSc3Z3V9vR0Tbn7Hb4Xjaxm5r4md1Bg87TAI</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><source>IEEE Xplore All Conference Series</source><creator>Surner, Martin ; Khelil, Abdelmajid</creator><creatorcontrib>Surner, Martin ; Khelil, Abdelmajid</creatorcontrib><description>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</description><identifier>EISSN: 2835-3420</identifier><identifier>EISBN: 9798350309294</identifier><identifier>DOI: 10.1109/SDS60720.2024.00018</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Accuracy ; Causality ; Correlation ; Domain Generalization ; Informed Machine Learning ; Knowledge engineering ; Neural networks ; Production ; Statistical learning ; Training</subject><ispartof>2024 11th IEEE Swiss Conference on Data Science (SDS), 2024, p.68-75</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10675979$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10675979$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Surner, Martin</creatorcontrib><creatorcontrib>Khelil, Abdelmajid</creatorcontrib><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><title>2024 11th IEEE Swiss Conference on Data Science (SDS)</title><addtitle>SDS</addtitle><description>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</description><subject>Accuracy</subject><subject>Causality</subject><subject>Correlation</subject><subject>Domain Generalization</subject><subject>Informed Machine Learning</subject><subject>Knowledge engineering</subject><subject>Neural networks</subject><subject>Production</subject><subject>Statistical learning</subject><subject>Training</subject><issn>2835-3420</issn><isbn>9798350309294</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2024</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotT9tKw0AUXAXBUvMF-rA_kHj2ls36prHVQIrQ6nM52T3RSJpK0gr9-y7oywwzDMMMY7cCMiHA3W-eNzlYCZkEqTMAEMUFS5x1hTKgwEmnL9lMRpUqLeGaJdP0HWNKCmF1MWOLslrV6fqBl3icsO9PvBra_bijwFfov7qBeE04Dt3wyZ9wivZ-4EvCw3EkvqaefnHwdMOuWuwnSv55zj6Wi_fyNa3fXqrysU47AfkhNarxgnyjPHojEQ21KpAN0kdU2hYmKB-HFY0wOlivQmt9aIUTULgcSc3Z3V9vR0Tbn7Hb4Xjaxm5r4md1Bg87TAI</recordid><startdate>20240530</startdate><enddate>20240530</enddate><creator>Surner, Martin</creator><creator>Khelil, Abdelmajid</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20240530</creationdate><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><author>Surner, Martin ; Khelil, Abdelmajid</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Causality</topic><topic>Correlation</topic><topic>Domain Generalization</topic><topic>Informed Machine Learning</topic><topic>Knowledge engineering</topic><topic>Neural networks</topic><topic>Production</topic><topic>Statistical learning</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Surner, Martin</creatorcontrib><creatorcontrib>Khelil, Abdelmajid</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Surner, Martin</au><au>Khelil, Abdelmajid</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</atitle><btitle>2024 11th IEEE Swiss Conference on Data Science (SDS)</btitle><stitle>SDS</stitle><date>2024-05-30</date><risdate>2024</risdate><spage>68</spage><epage>75</epage><pages>68-75</pages><eissn>2835-3420</eissn><eisbn>9798350309294</eisbn><coden>IEEPAD</coden><abstract>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</abstract><pub>IEEE</pub><doi>10.1109/SDS60720.2024.00018</doi><tpages>8</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 2835-3420 |
ispartof | 2024 11th IEEE Swiss Conference on Data Science (SDS), 2024, p.68-75 |
issn | 2835-3420 |
language | eng |
recordid | cdi_ieee_primary_10675979 |
source | IEEE Xplore All Conference Series |
subjects | Accuracy Causality Correlation Domain Generalization Informed Machine Learning Knowledge engineering Neural networks Production Statistical learning Training |
title | CIML-R: Causally Informed Machine Learning Based on Feature Relevance |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T07%3A57%3A09IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=CIML-R:%20Causally%20Informed%20Machine%20Learning%20Based%20on%20Feature%20Relevance&rft.btitle=2024%2011th%20IEEE%20Swiss%20Conference%20on%20Data%20Science%20(SDS)&rft.au=Surner,%20Martin&rft.date=2024-05-30&rft.spage=68&rft.epage=75&rft.pages=68-75&rft.eissn=2835-3420&rft.coden=IEEPAD&rft_id=info:doi/10.1109/SDS60720.2024.00018&rft.eisbn=9798350309294&rft_dat=%3Cieee_CHZPO%3E10675979%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10675979&rfr_iscdi=true |