Abstract
Decision tree is one of the most popular classifiers used in a wide range of real world problems for both prediction and classification (logic) rules discovery. A decision forest is an ensemble of decision trees, and it is often built for predicting class values more accurately than a single decision tree. Besides improving predictive performance, a decision forest can be seen as a pool of logic rules with great potential for knowledge discovery. However, an standard-sized decision forest usually generates a large number of logic rules that a user may not able to manage for effective knowledge analysis. In this paper, we propose a novel, problem (data set) independent framework for extracting those rules that are comparatively more accurate as well as reliable than others. We apply the proposed framework on rule sets generated from two different decision forest algorithms from a publicly available data set on dementia and compare the subsets of rules with the rules generated from a single J48 decision tree in order to show the effectiveness of the proposed framework.
Original language | English |
---|---|
Title of host publication | Proceedings of the Forteenth Australasian Data Mining Conference (AusDM 16) |
Place of Publication | Australia |
Publisher | CRPIT |
Pages | 1-8 |
Number of pages | 8 |
Publication status | Published - 2016 |
Event | The 14th Australasian Data Mining Conference: AusDM 2016 - Realm Hotel, Canberra, Australia Duration: 06 Dec 2016 → 08 Dec 2016 http://ausdm16.ausdm.org/ |
Conference
Conference | The 14th Australasian Data Mining Conference |
---|---|
Country/Territory | Australia |
City | Canberra |
Period | 06/12/16 → 08/12/16 |
Internet address |