Learning Explainable Decision Rules via Maximum Satisfiability

Loading...
Thumbnail Image

Access rights

openAccess
publishedVersion

URL

Journal Title

Journal ISSN

Volume Title

A1 Alkuperäisartikkeli tieteellisessä aikakauslehdessä

Date

2020

Major/Subject

Mcode

Degree programme

Language

en

Pages

6

Series

IEEE Access, Volume 8, pp. 218180-218185

Abstract

Decision trees are a popular choice for providing explainable machine learning, since they make explicit how different features contribute towards the prediction. We apply tools from constraint satisfaction to learn optimal decision trees in the form of sparse k-CNF (Conjunctive Normal Form) rules. We develop two methods offering different trade-offs between accuracy and computational complexity: one offline method that learns decision trees using the entire training dataset and one online method that learns decision trees over a local subset of the training dataset. This subset is obtained from training examples near a query point. The developed methods are applied on a number of datasets both in an online and an offline setting. We found that our methods learn decision trees which are significantly more accurate than those learned by existing heuristic approaches. However, the global decision tree model tends to be computationally more expensive compared to heuristic approaches. The online method is faster to train and finds smaller decision trees with an accuracy comparable to that of the k-nearest-neighbour method.

Description

Keywords

Ethics in modelling, Explainable AI, decision aiding tools, Satisfiability checking, machine learning

Other note

Citation

Cao, H, Sarlin, R & Jung, A 2020, ' Learning Explainable Decision Rules via Maximum Satisfiability ', IEEE Access, vol. 8, 9272729, pp. 218180-218185 . https://doi.org/10.1109/ACCESS.2020.3041040