194. Mixture of Experts (MoE)

A model architecture that combines multiple expert networks, each specialized in different tasks, to improve overall performance and scalability.

Last updated