Loading…

CIML-R: Causally Informed Machine Learning Based on Feature Relevance

Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domai...

Full description

Saved in:
Bibliographic Details
Main Authors: Surner, Martin, Khelil, Abdelmajid
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 75
container_issue
container_start_page 68
container_title
container_volume
creator Surner, Martin
Khelil, Abdelmajid
description Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.
doi_str_mv 10.1109/SDS60720.2024.00018
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_10675979</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10675979</ieee_id><sourcerecordid>10675979</sourcerecordid><originalsourceid>FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3</originalsourceid><addsrcrecordid>eNotT9tKw0AUXAXBUvMF-rA_kHj2ls36prHVQIrQ6nM52T3RSJpK0gr9-y7oywwzDMMMY7cCMiHA3W-eNzlYCZkEqTMAEMUFS5x1hTKgwEmnL9lMRpUqLeGaJdP0HWNKCmF1MWOLslrV6fqBl3icsO9PvBra_bijwFfov7qBeE04Dt3wyZ9wivZ-4EvCw3EkvqaefnHwdMOuWuwnSv55zj6Wi_fyNa3fXqrysU47AfkhNarxgnyjPHojEQ21KpAN0kdU2hYmKB-HFY0wOlivQmt9aIUTULgcSc3Z3V9vR0Tbn7Hb4Xjaxm5r4md1Bg87TAI</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><source>IEEE Xplore All Conference Series</source><creator>Surner, Martin ; Khelil, Abdelmajid</creator><creatorcontrib>Surner, Martin ; Khelil, Abdelmajid</creatorcontrib><description>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</description><identifier>EISSN: 2835-3420</identifier><identifier>EISBN: 9798350309294</identifier><identifier>DOI: 10.1109/SDS60720.2024.00018</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Accuracy ; Causality ; Correlation ; Domain Generalization ; Informed Machine Learning ; Knowledge engineering ; Neural networks ; Production ; Statistical learning ; Training</subject><ispartof>2024 11th IEEE Swiss Conference on Data Science (SDS), 2024, p.68-75</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10675979$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10675979$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Surner, Martin</creatorcontrib><creatorcontrib>Khelil, Abdelmajid</creatorcontrib><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><title>2024 11th IEEE Swiss Conference on Data Science (SDS)</title><addtitle>SDS</addtitle><description>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</description><subject>Accuracy</subject><subject>Causality</subject><subject>Correlation</subject><subject>Domain Generalization</subject><subject>Informed Machine Learning</subject><subject>Knowledge engineering</subject><subject>Neural networks</subject><subject>Production</subject><subject>Statistical learning</subject><subject>Training</subject><issn>2835-3420</issn><isbn>9798350309294</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2024</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotT9tKw0AUXAXBUvMF-rA_kHj2ls36prHVQIrQ6nM52T3RSJpK0gr9-y7oywwzDMMMY7cCMiHA3W-eNzlYCZkEqTMAEMUFS5x1hTKgwEmnL9lMRpUqLeGaJdP0HWNKCmF1MWOLslrV6fqBl3icsO9PvBra_bijwFfov7qBeE04Dt3wyZ9wivZ-4EvCw3EkvqaefnHwdMOuWuwnSv55zj6Wi_fyNa3fXqrysU47AfkhNarxgnyjPHojEQ21KpAN0kdU2hYmKB-HFY0wOlivQmt9aIUTULgcSc3Z3V9vR0Tbn7Hb4Xjaxm5r4md1Bg87TAI</recordid><startdate>20240530</startdate><enddate>20240530</enddate><creator>Surner, Martin</creator><creator>Khelil, Abdelmajid</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20240530</creationdate><title>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</title><author>Surner, Martin ; Khelil, Abdelmajid</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Causality</topic><topic>Correlation</topic><topic>Domain Generalization</topic><topic>Informed Machine Learning</topic><topic>Knowledge engineering</topic><topic>Neural networks</topic><topic>Production</topic><topic>Statistical learning</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Surner, Martin</creatorcontrib><creatorcontrib>Khelil, Abdelmajid</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Surner, Martin</au><au>Khelil, Abdelmajid</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>CIML-R: Causally Informed Machine Learning Based on Feature Relevance</atitle><btitle>2024 11th IEEE Swiss Conference on Data Science (SDS)</btitle><stitle>SDS</stitle><date>2024-05-30</date><risdate>2024</risdate><spage>68</spage><epage>75</epage><pages>68-75</pages><eissn>2835-3420</eissn><eisbn>9798350309294</eisbn><coden>IEEPAD</coden><abstract>Applications relying on machine learning and statistical learning techniques, such as neural networks, have significantly grown over the past decade. Nevertheless, as these techniques learn only from observational data, they suffer from spurious correlations that may limit their performance in domain shifts. In this paper, we address this issue. We propose an approach that guides neural networks during the training phase using explanations. User-defined causal knowledge is incorporated to guide the considered neural network. Such guidance reduces the risk of spurious correlations, allowing for a focus on the causal structure. Our method extends the feed-forward step of neural networks to track the feature relevance. The resulting feature relevance is accounted for in a regularization term in the loss function. We compare the performance of a conventional neural network with and without the proposed extension. This comparison indicates that the learning process successfully adopts the causal mechanism, while the conventional approach learns based on spurious correlations.</abstract><pub>IEEE</pub><doi>10.1109/SDS60720.2024.00018</doi><tpages>8</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2835-3420
ispartof 2024 11th IEEE Swiss Conference on Data Science (SDS), 2024, p.68-75
issn 2835-3420
language eng
recordid cdi_ieee_primary_10675979
source IEEE Xplore All Conference Series
subjects Accuracy
Causality
Correlation
Domain Generalization
Informed Machine Learning
Knowledge engineering
Neural networks
Production
Statistical learning
Training
title CIML-R: Causally Informed Machine Learning Based on Feature Relevance
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T07%3A57%3A09IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=CIML-R:%20Causally%20Informed%20Machine%20Learning%20Based%20on%20Feature%20Relevance&rft.btitle=2024%2011th%20IEEE%20Swiss%20Conference%20on%20Data%20Science%20(SDS)&rft.au=Surner,%20Martin&rft.date=2024-05-30&rft.spage=68&rft.epage=75&rft.pages=68-75&rft.eissn=2835-3420&rft.coden=IEEPAD&rft_id=info:doi/10.1109/SDS60720.2024.00018&rft.eisbn=9798350309294&rft_dat=%3Cieee_CHZPO%3E10675979%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i106t-53bc1ecb3cac52aa5ef3de7d2cde734785d3c1178b154d7c3df7cdf1910896ae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10675979&rfr_iscdi=true