Previous work on feature weighting for case-based learning
algorithms has tended to use either global weights or weights that vary
over extremely local regions of the case space. This paper examines the
use of coarsely local weighting schemes, where feature weights are allowed
to vary but are identical for groups or clusters of cases. We present a new
technique, called class distribution weighting (CDW), that allows weights
to vary at the class level. We further extend CDW into a family of related
techniques that exhibit varying degrees of locality, from global to local.
The class distribution techniques are then applied to a set of eleven
concept learning tasks. We nd that one or more of the CDW variants
signicantly improves classication accuracy for nine of the eleven tasks.
In addition, we nd that the relative importance of classes, features, and
feature values in a particular domain determines which variant is most
successful