Abstract
The problem of optimal data fusion in multiple detector systems is studied in the case where training examples are available, but no a priori information is available about the probability distributions of errors committed by the individual detectors. Earlier solutions to this problem require some knowledge of the error distributions of the detectors, for example, either in a parametric form or in a closed analytical form. Here we show that, given a sufficiently large training sample, an optimal fusion rule can be implemented with an arbitrary level of confidence. We first consider the classical cases of Bayesian rule and Neyman-Pearson test for a system of independent detectors. Then we show a general result that any test function with a suitable Lipschitz property can be implemented with arbitrary precision, based on a training sample whose size is a function of the Lipschitz constant, number of parameters, and empirical measures. The general case subsumes the cases of non-independent and correlated detectors.
Original language | English |
---|---|
Pages | 697-704 |
Number of pages | 8 |
State | Published - 1996 |
Event | Proceedings of the 1996 IEEE/SICE/RSJ International Conference on Multisensor Fusion and Integration for Intelligent Systems - Washington, DC, USA Duration: Dec 8 1996 → Dec 11 1996 |
Conference
Conference | Proceedings of the 1996 IEEE/SICE/RSJ International Conference on Multisensor Fusion and Integration for Intelligent Systems |
---|---|
City | Washington, DC, USA |
Period | 12/8/96 → 12/11/96 |