“School of Cognitive”
Back to Papers HomeBack to Papers of School of Cognitive
Paper IPM / Cognitive / 11608 |
|
||||||||||
Abstract: | |||||||||||
Mixture of Experts (ME) is a modular neural network architecture for supervised learning. In this paper, we propose an evidence-based ME to deal with the classification problem. In the basic form of ME the problem space is automatically divided into several subspaces for the experts and the outputs of experts are combined
by a gating network. Satisfactory performance of the basic ME depends on thediversity among experts. In conventional ME, different initialization of experts and supervision of the gating network during the learning procedure , provide the diversity. The main idea of our proposed method is to employ the Dempster-Shafer (D-S) theory of evidence to improve determination of learning parameters (which results more diverse experts) and the way of combining experts' decisions.Experimental results with some data sets from uci epository show that our proposed method yields better classification rates as compared to basic ME and static combining of neural network based on D-S theory.
Download TeX format |
|||||||||||
back to top |