The Community for Technology Leaders
Green Image
ISSN: 0162-8828
Jian-Huang Lai , Sun Yat-Sen University, Guangzhou
Pong C. Yuen , Hong Kong Baptist University, Kowloon
Andy J. Ma , Hong Kong Baptist University, Hong Kong
This paper addresses the independent assumption issue in fusion process. In the last decade, dependency modeling techniques were developed under a specific distribution of classifiers or by estimating the joint distribution of the posterior. This paper proposes a new framework to model the dependency between features without any assumption on feature/classifier distribution, and overcome the difficulty in estimating the high-dimensional joint distribution. In this paper, we prove that feature dependency can be modeled by a linear combination of the posterior probabilities under some mild assumptions. Based on the linear combination property, two methods, namely Linear Classifier Dependency Modeling (LCDM) and Linear Feature Dependency Modeling (LFDM), are derived and developed for dependency modeling in classifier level and feature level, respectively. The optimal models for LCDM and LFDM are learned by maximizing the margin between the genuine and imposter posterior probabilities. Both synthetic data and real datasets are used for experiments. Experimental results show that LCDM and LFDM with dependency modeling outperform existing classifier level and feature level combination methods under non-normal distributions and on four real databases, respectively. Comparing the classifier level and feature level fusion methods, LFDM gives the best performance.
Joints, Mathematical model, Computational modeling, Kernel, Vectors, Linear programming, Optimization, Classifier design and evaluation, Computing Methodologies, Pattern Recognition, Design Methodology
Jian-Huang Lai, Pong C. Yuen, Andy J. Ma, "Linear Dependency Modeling for Classifier Fusion and Feature Combination", IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. , no. , pp. 0, 5555, doi:10.1109/TPAMI.2012.198
264 ms
(Ver 3.3 (11022016))