An evaluation of heuristics for rule ranking

Stephan Dreiseitl, Melanie Osl, Christian Baumgartner, Staal Vinterbo

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)

Abstract

Objective: To evaluate and compare the performance of different rule-ranking algorithms for rule-based classifiers on biomedical datasets. Methodology: Empirical evaluation of five rule ranking algorithms on two biomedical datasets, with performance evaluation based on ROC analysis and 5×2 cross-validation. Results: On a lung cancer dataset, the area under the ROC curve (AUC) of, on average, 14267.1 rules was 0.862. Multi-rule ranking found 13.3 rules with an AUC of 0.852. Four single-rule ranking algorithms, using the same number of rules, achieved average AUC values of 0.830, 0.823, 0.823, and 0.822, respectively. On a prostate cancer dataset, an average of 339265.3 rules had an AUC of 0.934, while 9.4 rules obtained from multi-rule and single-rule rankings had average AUCs of 0.932, 0.926, 0.925, 0.902 and 0.902, respectively. Conclusion: Multi-variate rule ranking performs better than the single-rule ranking algorithms. Both single-rule and multi-rule methods are able to substantially reduce the number of rules while keeping classification performance at a level comparable to the full rule set.

Original languageEnglish
Pages (from-to)175-180
Number of pages6
JournalArtificial Intelligence in Medicine
Volume50
Issue number3
DOIs
Publication statusPublished - Nov 2010

Keywords

  • Lung cancer
  • Prostate cancer
  • Rule evaluation metrics
  • Rule ranking
  • Breast Neoplasms/pathology
  • Area Under Curve
  • Lung Neoplasms/pathology
  • Artificial Intelligence
  • Humans
  • Male
  • Algorithms
  • Prostatic Neoplasms/pathology
  • Female

Fingerprint

Dive into the research topics of 'An evaluation of heuristics for rule ranking'. Together they form a unique fingerprint.

Cite this