Machine Learning with Expert Models: A Primer

<p>Expert models are one of the most useful inventions in Machine Learning, yet they hardly receive as much attention as they deserve. In fact, expert modeling does not only allow us to train neural networks that are &ldquo;outrageously large&rdquo; (more on that later), they also allow us to build models that learn more like the human brain, that is, different regions specialize in different types of input.</p> <p>In this article, we&rsquo;ll take a tour of the key innovations in expert modeling which ultimately lead to recent breakthroughs such as the Switch Transformer and the Expert Choice Routing algorithm. But let&rsquo;s go back first to the paper that started it all: &ldquo;Mixtures of Experts&rdquo;.</p> <h2>Mixtures of Experts (1991)</h2> <p><img alt="" src="https://miro.medium.com/v2/resize:fit:499/1*Nu6cVmRFzrvR2EVJ4gNNaA.png" style="height:444px; width:499px" /></p> <p>The original MoE model from 1991. Image credit:&nbsp;<a href="https://www.cs.toronto.edu/~hinton/absps/jjnh91.pdf" rel="noopener ugc nofollow" target="_blank">Jabocs et al 1991, Adaptive Mixtures of Local Experts</a>.</p> <p>The idea of mixtures of experts (MoE) traces back more than 3 decades ago, to a 1991&nbsp;<a href="https://www.cs.toronto.edu/~hinton/absps/jjnh91.pdf" rel="noopener ugc nofollow" target="_blank">paper</a>&nbsp;co-authored by none other than the godfather of AI, Geoffrey Hinton. The key idea in MoE is to model an output &ldquo;y&rdquo; by combining a number of &ldquo;experts&rdquo; E, the weight of each is being controlled by a &ldquo;gating network&rdquo; G:</p> <p>An expert in this context can be any kind of model, but is usually chosen to be a multi-layered neural network, and the gating network is</p> <p><a href="https://towardsdatascience.com/machine-learning-with-expert-models-a-primer-6c74585f223f"><strong>Learn More</strong></a></p>