Proceedings of 16th International Conference on Data Engineering (Cat. No.00CB37073) (2000)
San Diego, California
Feb. 28, 2000 to Mar. 3, 2000
Bernd-Uwe Pagel , SAP AG
Flip Korn , AT&T Labs-Research
Christos Faloutsos , Carnegie Mellon University
Nearest neighbor queries are important in many settings, including spatial databases (Find the k closest cities) and multimedia databases (Find the k most similar images). Previous analyses have concluded that nearest neighbor search is hopeless in high dimensions, due to the notorious "curse of dimensionality". However, their precise analysis over real data sets is still an open problem.The typical and often implicit assumption in previous studies is that the data is uniformly distributed, with independence between attributes. However, real data sets overwhelmingly disobey these assumptions; rather, they typically are skewed and exhibit intrinsic ("fractal") dimensionalities that are much lower than their embedding dimension, e.g., due to subtle dependencies between attributes.In this paper, we show how the Hausdorff and Correlation fractal dimensions of a data set can yield extremely accurate formulas that can predict I/O performance to within one standard deviation. The practical contributions of this work are our accurate formulas which can be used for query optimization in spatial and multimedia databases. The theoretical contribution is the 'deflation' of the dimensionality curse.Our theoretical and empirical results show that previous worst-case analyses of nearest neighbor search in high dimensions are over-pessimistic, to the point of being unrealistic. The performance depends critically on the intrinsic ("fractal") dimensionality as opposed to the embedding dimension that the uniformity assumption incorrectly implies.
F. Korn, C. Faloutsos and B. Pagel, "Deflating the Dimensionality Curse Using Multiple Fractal Dimensions," Proceedings of 16th International Conference on Data Engineering (Cat. No.00CB37073)(ICDE), San Diego, California, 2000, pp. 589.