2021
DOI: 10.1609/aaai.v35i5.16555
|View full text |Cite
|
Sign up to set email alerts
|

Learning Accurate and Interpretable Decision Rule Sets from Neural Networks

Abstract: This paper proposes a new paradigm for learning a set of independent logical rules in disjunctive normal form as an interpretable model for classification. We consider the problem of learning an interpretable decision rule set as training a neural network in a specific, yet very simple two-layer architecture. Each neuron in the first layer directly maps to an interpretable if-then rule after training, and the output neuron in the second layer directly maps to a disjunction of the first layer rules to form the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 17 publications
(17 citation statements)
references
References 16 publications
0
17
0
Order By: Relevance
“…where 𝜎 (•) denotes the Softmax function, đť‘  denotes the SIS vector (17), đť‘  ′ is the normalized and scaled SIS with factor đť‘Ž, and đť‘  ′ đť‘ť is the đť‘ťth element of đť‘  ′ for feature variable đť‘Ź đť‘ť . In this step, đť‘š WCS IPs are created, each by sampling |P ′𝑒 đť‘– | = 𝜌 |P đť‘’ | (rounded, 0 < 𝜌 < 1) feature variables according to the probability Prob đť‘ť (18), and |N đť‘– | = đť‘› wcs observations uniformly from the observations set N .…”
Section: Weighted Column Sampling Optimizationmentioning
confidence: 99%
See 2 more Smart Citations
“…where 𝜎 (•) denotes the Softmax function, đť‘  denotes the SIS vector (17), đť‘  ′ is the normalized and scaled SIS with factor đť‘Ž, and đť‘  ′ đť‘ť is the đť‘ťth element of đť‘  ′ for feature variable đť‘Ź đť‘ť . In this step, đť‘š WCS IPs are created, each by sampling |P ′𝑒 đť‘– | = 𝜌 |P đť‘’ | (rounded, 0 < 𝜌 < 1) feature variables according to the probability Prob đť‘ť (18), and |N đť‘– | = đť‘› wcs observations uniformly from the observations set N .…”
Section: Weighted Column Sampling Optimizationmentioning
confidence: 99%
“…Note that these varying implementations of branch-and-bound solution procedure only affect the time efficiencies in solving the models, and do not change the solution values. The proposed SIS-based WCS optimization method samples feature variables according to their SIS values' associated Softmax probabilities (18). As an alternative for comparison, the SIS-based feature sampling approach in WCS can be replaced with a uniform sampling approach, called random column sampling (RCS), which associates equal sampling probabilities for each feature variable.…”
Section: Branch-and-bound Variantsmentioning
confidence: 99%
See 1 more Smart Citation
“…The optimization techniques in AMIE speed up the rule mining procedure by one order but heavily sacrifice the expressiveness of Horn rules. Neural networks are also adopted for learning logic rules in some recent work 34 …”
Section: Related Workmentioning
confidence: 99%
“…Neural networks are also adopted for learning logic rules in some recent work. 34 Association patterns are less expressive compared to Horn rules. For example, the patterns induced from LLC can be expressed as the following Horn rules:…”
Section: Association Pattern Mining and Logic Rule Miningmentioning
confidence: 99%