Investigating the impact of independent rule fitnesses in a learning classifier system

  • Achieving at least some level of explainability requires complex analyses for many machine learning systems, such as common black-box models. We recently proposed a new rule-based learning system, SupRB, to construct compact, interpretable and transparent models by utilizing separate optimizers for the model selection tasks concerning rule discovery and rule set composition. This allows users to specifically tailor their model structure to fulfil use-case specific explainability requirements. From an optimization perspective, this allows us to define clearer goals and we find that—in contrast to many state of the art systems—this allows us to keep rule fitnesses independent. In this paper we investigate this system’s performance thoroughly on a set of regression problems and compare it against XCSF, a prominent rule-based learning system. We find the overall results of SupRB’s evaluation comparable to XCSF’s while allowing easier control of model structure and showing a substantiallyAchieving at least some level of explainability requires complex analyses for many machine learning systems, such as common black-box models. We recently proposed a new rule-based learning system, SupRB, to construct compact, interpretable and transparent models by utilizing separate optimizers for the model selection tasks concerning rule discovery and rule set composition. This allows users to specifically tailor their model structure to fulfil use-case specific explainability requirements. From an optimization perspective, this allows us to define clearer goals and we find that—in contrast to many state of the art systems—this allows us to keep rule fitnesses independent. In this paper we investigate this system’s performance thoroughly on a set of regression problems and compare it against XCSF, a prominent rule-based learning system. We find the overall results of SupRB’s evaluation comparable to XCSF’s while allowing easier control of model structure and showing a substantially smaller sensitivity to random seeds and data splits. This increased control can aid in subsequently providing explanations for both training and final structure of the model.show moreshow less

Download full text files

Export metadata

Statistics

Number of document requests

Additional Services

Share in Twitter Search Google Scholar
Metadaten
Author:Michael HeiderORCiDGND, Helena StegherrORCiDGND, Jonathan WurthORCiDGND, Roman Sraj, Jörg HähnerORCiDGND
URN:urn:nbn:de:bvb:384-opus4-1039736
Frontdoor URLhttps://opus.bibliothek.uni-augsburg.de/opus4/103973
ISBN:9783031210938OPAC
ISBN:9783031210945OPAC
ISSN:0302-9743OPAC
ISSN:1611-3349OPAC
Parent Title (English):Lecture Notes in Computer Science
Publisher:Springer
Place of publication:Cham
Type:Article
Language:English
Year of first Publication:2022
Publishing Institution:Universität Augsburg
Release Date:2023/04/25
Volume:13627
First Page:142
Last Page:156
Note:
Published in: Bioinspired Optimization Methods and Their Applications: 10th International Conference, BIOMA 2022, Maribor, Slovenia, November 17–18, 2022, Proceedings. Edited by Marjan Mernik, Tome Eftimov, Matej Črepinšek
DOI:https://doi.org/10.1007/978-3-031-21094-5_11
Institutes:Fakultät für Angewandte Informatik
Fakultät für Angewandte Informatik / Institut für Informatik
Fakultät für Angewandte Informatik / Institut für Informatik / Lehrstuhl für Organic Computing
Dewey Decimal Classification:0 Informatik, Informationswissenschaft, allgemeine Werke / 00 Informatik, Wissen, Systeme / 004 Datenverarbeitung; Informatik
Licence (German):Deutsches Urheberrecht