KNOWLEDGE FUSION FOR PROBABILISTIC GENERATIVE CLASSIFIERS WITH DATA MINING APPLICATIONS
₹4,500.00
10000 in stock
SupportDescription
ABSTRACT
If knowledge such as classification rules are extracted from sample data in a distributed way, it may be necessary to combine or fuse these rules. In a conventional approach this would typically bed one either by combining the classifiers’ outputs (e.g., in form of a classifier ensemble) or by combining the sets of classification rules (e.g. ,by weighting them individually). In this article, we introduce a new way of fusing classifiers at the level of parameters of classification rules. This technique is based on the use of probabilistic generative classifiers using multinomial distributions for categorical input dimensions and multivariate normal distributions for the continuous ones. That means, we have distributions such as Dirichlet or normal-Wishart distributions over parameters of the classifier. We refer to these distributions as hyper distributions or second-order distributions. We show that fusing two (or more) classifiers can be done by multiplying the hyper-distributions of the parameters and derive simple formulas for that task. Properties of this new approach are demonstrated with a few experiments. The main advantage of this fusion approach is that the hyper-distributions are retained throughout the fusion process. Thus, the fused components may, e.g., be used in subsequent training steps (online training).


