Abstract
It is well-known that a combination of many different predictors can improve predictions. This combination is usually effected by majority in classification or by simple averaging in regression, but one can also use a weighted combination of the networks. The first section of this chapter summarises the main ideas of a recent study by Krogh and Vedelsby on network committees for simple interpolation tasks. The generalisation performance of the committee is seen to depend on the variation of the output of ensemble members, which the authors refer to as the ambiguity. The second section generalises the results to the prediction of conditional probability densities. A similar dependence on an ambiguity term is found, which suggests that diversity in the committee is a crucial requirement for improving the generalisation performance. The third section discusses aspects of the weighting scheme. The optimisation of the weights on the basis of the training data is shown to lead to a sub-optimal behaviour, and the evidence scheme is suggested as an approach that arises naturally within a Bayesian framework. However, numerical deficiencies of the latter are pointed out, and the chapter concludes with the introduction of an alternative scheme.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this chapter
Cite this chapter
Husmeier, D. (1999). Network Committees and Weighting Schemes. In: Neural Networks for Conditional Probability Estimation. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0847-4_13
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0847-4_13
Publisher Name: Springer, London
Print ISBN: 978-1-85233-095-8
Online ISBN: 978-1-4471-0847-4
eBook Packages: Springer Book Archive