Not Logged In

Concept Learning and the Problem of Small Disjuncts

Ideally, definitions induced from examples should consist of all, and only, disjuncts that are meaningful (e.g., as measured by a statistical significance test) and have a low error rate. Existing inductive systems create definitions that are ideal with regard to large disjuncts, but far from ideal with regard to small disjuncts, where a small (large) disjunct is one that correctly classifies few (many) training examples. The problem with small disjuncts is that many of them have high rates of misclassification, and it is difficult to eliminate the error­prone small disjuncts from a definition without adversely affecting other disjuncts in the definition. Various approaches to this problem are evaluated, including the novel approach of using a bias different than the ``maximum generality'' bias. This approach, and some others, prove partly successful, but the problem of small disjuncts remains open.

Citation

R. Holte, L. Acker, B. Porter. "Concept Learning and the Problem of Small Disjuncts". International Joint Conference on Artificial Intelligence (IJCAI), Detroit, Michigan, USA, pp 813-818, August 1989.

Keywords: disjuncts, machine learning
Category: In Conference

BibTeX

@incollection{Holte+al:IJCAI89,
  author = {Robert Holte and Liane E. Acker and Bruce W. Porter},
  title = {Concept Learning and the Problem of Small Disjuncts},
  Pages = {813-818},
  booktitle = {International Joint Conference on Artificial Intelligence
    (IJCAI)},
  year = 1989,
}

Last Updated: June 18, 2007
Submitted by Staurt H. Johnson

University of Alberta Logo AICML Logo