Department of Computer Science
Reduced analytic dependency modeling: Robust fusion for visual recognition
This paper addresses the robustness issue of information fusion for visual recognition. Analyzing limitations in existing fusion methods, we discover two key factors affecting the performance and robustness of a fusion model under different data distributions, namely (1) data dependency and (2) fusion assumption on posterior distribution. Considering these two factors, we develop a new framework to model dependency based on probabilistic properties of posteriors without any assumption on the data distribution. Making use of the range characteristics of posteriors, the fusion model is formulated as an analytic function multiplied by a constant with respect to the class label. With the analytic fusion model, we give an equivalent condition to the independent assumption and derive the dependency model from the marginal distribution property. Since the number of terms in the dependency model increases exponentially, the Reduced Analytic Dependency Model (RADM) is proposed based on the convergent property of analytic function. Finally, the optimal coefficients in the RADM are learned by incorporating label information from training data to minimize the empirical classification error under regularized least square criterion, which ensures the discriminative power. Experimental results from robust non-parametric statistical tests show that the proposed RADM method statistically significantly outperforms eight state-of-the-art score-level fusion methods on eight image/video datasets for different tasks of digit, flower, face, human action, object, and consumer video recognition. © 2014 Springer Science+Business Media New York.
Dependency modeling, Probabilistic constraints, Robustness, Score-level fusion, Visual recognition
Source Publication Title
International Journal of Computer Vision
Link to Publisher's Edition
Ma, Andy J., and Pong C. Yuen. "Reduced analytic dependency modeling: Robust fusion for visual recognition." International Journal of Computer Vision 109.3 (2014): 233-251.