Volume 8, Issue 4
OVERVIEW

Practical and theoretical aspects of mixture‐of‐experts modeling: An overview

Hien D. Nguyen

Corresponding Author

E-mail address: h.nguyen5@latrobe.edu.au

Department of Mathematics and Statistics, La Trobe University, 3086, Bundoora, Victoria, Australia

Correspondence

Hien D. Nguyen, Department of Mathematics and Statistics, La Trobe University, 3086 Bundoora, Victoria, Australia.

Email: h.nguyen5@latrobe.edu.au

Search for more papers by this author
Faicel Chamroukhi

Normandie Universite, UNICAEN, CNRS, Laboratoire de Mathematiques Nicolas Oresme – LMNO, 14000, Caen, France

Search for more papers by this author
First published: 13 February 2018
Citations: 2

Abstract

Mixture‐of‐experts (MoE) models are a powerful paradigm for modeling data arising from complex data generating processes (DGPs). In this article, we demonstrate how different MoE models can be constructed to approximate the underlying DGPs of arbitrary types of data. Due to the probabilistic nature of MoE models, we propose the maximum quasi‐likelihood (MQL) approach as a method for estimating MoE model parameters from data, and we provide conditions under which MQL estimators are consistent and asymptotically normal. The blockwise minorization–maximization (blockwise‐MM) algorithm framework is proposed as an all‐purpose method for constructing algorithms for obtaining MQL estimators. An example derivation of a blockwise‐MM algorithm is provided. We then present a method for constructing information criteria for estimating the number of components in MoE models and provide justification for the classic Bayesian information criterion (BIC). We explain how MoE models can be used to conduct classification, clustering, and regression and illustrate these applications via two worked examples.

This article is categorized under:

  • Algorithmic Development > Statistics
  • Technologies > Structure Discovery and Clustering

Abstract

Schematic diagram of the neural network architecture of a mixture of experts model.

Number of times cited according to CrossRef: 2

  • Model‐based clustering and classification of functional data, WIREs Data Mining and Knowledge Discovery , 10.1002/widm.1298, 9, 4, (2019).
  • Regularized Estimation and Feature Selection in Mixtures of Gaussian-Gated Experts Models, Statistics and Data Science, 10.1007/978-981-15-1960-4_3, (42-56), (2019).

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.