The Community for Technology Leaders
RSS Icon
Subscribe
Issue No.04 - April (1991 vol.13)
pp: 340-354
ABSTRACT
<p>An iterative algorithm that finds a locally optimal partition for an arbitrary loss function, in time linear in N for each iteration is presented. The algorithm is a K-means-like clustering algorithm that uses as its distance measure a generalization of Kullback's information divergence. Moreover, it is proven that the globally optimal partition must satisfy a nearest neighbour condition using divergence as the distance measure. These results generalize similar results of L. Breiman et al. (1984) to an arbitrary number of classes or regression variables and to an arbitrary number of bills. Experimental results on a text-to-speech example are provided and additional applications of the algorithm, including the design of variable combinations, surrogate splits, composite nodes, and decision graphs, are suggested.</p>
INDEX TERMS
speech recognition; partitioning; regression trees; iterative algorithm; clustering algorithm; Kullback's information divergence; text-to-speech; surrogate splits; composite nodes; decision graphs; decision theory; iterative methods; speech recognition; trees (mathematics)
CITATION
P.A. Chou, "Optimal Partitioning for Classification and Regression Trees", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol.13, no. 4, pp. 340-354, April 1991, doi:10.1109/34.88569
16 ms
(Ver 2.0)

Marketing Automation Platform Marketing Automation Tool