Affiliation:
1. Department of Brain and Cognitive Sciences, University of Rochester, Rochester, NY 14627 USA
Abstract
This article investigates the bias and variance of mixtures-of-experts (ME) architectures. The variance of an ME architecture can be expressed as the sum of two terms: the first term is related to the variances of the expert networks that comprise the architecture and the second term is related to the expert networks' covariances. One goal of this article is to study and quantify a number of properties of ME architectures via the metrics of bias and variance. A second goal is to clarify the relationships between this class of systems and other systems that have recently been proposed. It is shown that in contrast to systems that produce unbiased experts whose estimation errors are uncorrelated, ME architectures produce biased experts whose estimates are negatively correlated.
Subject
Cognitive Neuroscience,Arts and Humanities (miscellaneous)
Cited by
52 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. COMET: Learning Cardinality Constrained Mixture of Experts with Trees and Local Search;Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining;2023-08-04
2. ME-SFP: A Mixture-of-Experts-Based Approach for Software Fault Prediction;IEEE Transactions on Reliability;2023
3. Mixture of Experts for Facial Forgery Detection;Journal of Imaging Science and Technology;2022-11-01
4. ST-ExpertNet: A Deep Expert Framework for Traffic Prediction;IEEE Transactions on Knowledge and Data Engineering;2022
5. Semi-explicit mixture of experts based on information table;Journal of Ambient Intelligence and Humanized Computing;2021-11-30