dc.contributor.author | Saarela, Mirka | |
dc.contributor.author | Jauhiainen, Susanne | |
dc.date.accessioned | 2021-02-08T09:36:50Z | |
dc.date.available | 2021-02-08T09:36:50Z | |
dc.date.issued | 2021 | |
dc.identifier.citation | Saarela, M., & Jauhiainen, S. (2021). Comparison of feature importance measures as explanations for classification models. <i>SN Applied Sciences</i>, <i>3</i>(2), Article 272. <a href="https://doi.org/10.1007/s42452-021-04148-9" target="_blank">https://doi.org/10.1007/s42452-021-04148-9</a> | |
dc.identifier.other | CONVID_51361793 | |
dc.identifier.uri | https://jyx.jyu.fi/handle/123456789/74012 | |
dc.description.abstract | Explainable artificial intelligence is an emerging research direction helping the user or developer of machine learning models understand why models behave the way they do. The most popular explanation technique is feature importance. However, there are several different approaches how feature importances are being measured, most notably global and local. In this study we compare different feature importance measures using both linear (logistic regression with L1 penalization) and non-linear (random forest) methods and local interpretable model-agnostic explanations on top of them. These methods are applied to two datasets from the medical domain, the openly available breast cancer data from the UCI Archive and a recently collected running injury data. Our results show that the most important features differ depending on the technique. We argue that a combination of several explanation techniques could provide more reliable and trustworthy results. In particular, local explanations should be used in the most critical cases such as false negatives. | en |
dc.format.mimetype | application/pdf | |
dc.language | eng | |
dc.language.iso | eng | |
dc.publisher | Springer | |
dc.relation.ispartofseries | SN Applied Sciences | |
dc.rights | CC BY 4.0 | |
dc.subject.other | feature importance | |
dc.subject.other | explainable artificial intelligence | |
dc.subject.other | interpretable models | |
dc.subject.other | random forest | |
dc.subject.other | logistic regression | |
dc.title | Comparison of feature importance measures as explanations for classification models | |
dc.type | article | |
dc.identifier.urn | URN:NBN:fi:jyu-202102081454 | |
dc.contributor.laitos | Informaatioteknologian tiedekunta | fi |
dc.contributor.laitos | Faculty of Information Technology | en |
dc.contributor.oppiaine | Tietotekniikka | fi |
dc.contributor.oppiaine | Mathematical Information Technology | en |
dc.type.uri | http://purl.org/eprint/type/JournalArticle | |
dc.type.coar | http://purl.org/coar/resource_type/c_2df8fbb1 | |
dc.description.reviewstatus | peerReviewed | |
dc.relation.issn | 2523-3963 | |
dc.relation.numberinseries | 2 | |
dc.relation.volume | 3 | |
dc.type.version | publishedVersion | |
dc.rights.copyright | © 2021 the Authors | |
dc.rights.accesslevel | openAccess | fi |
dc.relation.grantnumber | 311877 | |
dc.subject.yso | tekoäly | |
dc.subject.yso | koneoppiminen | |
dc.subject.yso | luokitus (toiminta) | |
dc.format.content | fulltext | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p2616 | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p21846 | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p12668 | |
dc.rights.url | https://creativecommons.org/licenses/by/4.0/ | |
dc.relation.doi | 10.1007/s42452-021-04148-9 | |
dc.relation.funder | Research Council of Finland | en |
dc.relation.funder | Suomen Akatemia | fi |
jyx.fundingprogram | Research profiles, AoF | en |
jyx.fundingprogram | Profilointi, SA | fi |
jyx.fundinginformation | This research was supported by the Academy of Finland (Grant No. 311877) and is related to the thematic research area DEMO (Decision Analytics Utilizing Causal Models and Multiobjective Optimization, jyu.fi/demo) of the University of Jyväskylä, Finland. | |
dc.type.okm | A1 | |