In pattern recognition and image processing. the major application areas of
cluster analysis, human eyes seem to possess a singular aptitude to group
objects and find important structures in an efficient and effective way. Th
us, a clustering algorithm simulating a visual system may solve some basic
problems in these areas of research. From this point of view, we propose a
new approach to data clustering by modeling the blurring effect of lateral
retinal interconnections based on scale space theory. In this approach, a d
ata set is considered as an image with each light point located at a datum
position. As we blur this image, smaller light blobs merge into larger ones
until the whole image becomes one light blob at a low enough level of reso
lution. By identifying each blob with a cluster, the blurring process gener
ates a family of clusterings along the hierarchy. The advantages of the pro
posed approach are: 1) The derived algorithms are computationally stable an
d insensitive to initialization and they are totally free from solving diff
icult global optimization problems. 2) It facilitates the construction of n
ew checks on cluster validity and provides the final clustering a significa
nt degree of robustness to noise in data and change in scale. 3) It is more
robust in cases where hyperellipsoidal partitions may not be assumed. 4) I
t is suitable for the task of preserving the structure and integrity of the
outliers in the clustering process. 5) The clustering is highly consistent
with that perceived by human eyes. 6) The new approach provides a unified
framework for scale-related clustering algorithms recently derived from man
y different fields such as estimation theory, recurrent signal processing o
n selforganization feature maps, information theory and statistical mechani
cs, and radial basis function neural networks.