AN EXPERIMENTAL COMPARISON OF THE NEAREST-NEIGHBOR AND NEAREST-HYPERRECTANGLE ALGORITHMNS

Citation
D. Wettschereck et Tg. Dietterich, AN EXPERIMENTAL COMPARISON OF THE NEAREST-NEIGHBOR AND NEAREST-HYPERRECTANGLE ALGORITHMNS, Machine learning, 19(1), 1995, pp. 5-27
Citations number
13
Categorie Soggetti
Computer Sciences","Computer Science Artificial Intelligence",Neurosciences
Journal title
ISSN journal
08856125
Volume
19
Issue
1
Year of publication
1995
Pages
5 - 27
Database
ISI
SICI code
0885-6125(1995)19:1<5:AECOTN>2.0.ZU;2-H
Abstract
Algorithms based on Nested Generalized Exemplar (NGE) theory (Salzberg , 1991) classify new data points by computing their distance to the ne arest ''generalized exemplar'' (i.e., either a point or an axis-parall el rectangle). They combine the distance-based character of nearest ne ighbor (NN) classifiers with the axis-parallel rectangle representatio n employed in many rule-learning systems. An implementation of NGE was compared to the k-nearest neighbor (kNN) algorithm in 11 domains and found to be significantly inferior to kNN in 9 of them Several modific ations of NGE were studied to understand the cause of its poor perform ance. These show that its performance can be substantially improved by preventing NGE from creating overlapping rectangles, while still allo wing complete nesting of rectangles. Performance can be further improv ed by modifying the distance metric to allow weights on each of the fe atures (Salzberg, 1991). Best results were obtained in this study when the weights were computed using mutual information between the featur es and the output class. The best version of NGE developed is a batch algorithm (BNGE FWMI) that has no user-tunable parameters. BNGE FWMI's performance is comparable to the first-nearest neighbor algorithm (al so incorporating feature weights). However, the k-nearest neighbor alg orithm is still significantly superior to BNGE FWMI in 7 of the 11 dom ains, and inferior to it in only 2. We conclude that, even with our im provements, the NGE approach is very sensitive to the shape of the dec ision boundaries in classification problems. in domains where the deci sion boundaries are axis-parallel, the NGE approach can produce excell ent generalization with interpretable hypotheses. In all domains teste d, NGE algorithms require much less memory to store generalized exempl ars than is required by NN algorithms.