Modularizing Language Models with Composition of Experts

Saturday 01 February 2025


The quest for a more efficient and effective way to train large language models (LLMs) has led researchers to explore new approaches, including the composition of experts. This modular compound AI system leverages multiple LLMs to achieve better performance on various tasks.


The idea behind this approach is simple yet powerful: instead of relying on a single, massive model, you can combine smaller, specialized models to tackle specific challenges. By doing so, you can reduce the complexity and computational requirements of training and deploying these models.


The researchers behind this work used a combination of existing datasets and synthetic prompts to create a diverse set of language tasks. These tasks were then fed into a variety of LLMs, including some that are widely known for their performance on specific domains, such as biology and law.


To evaluate the effectiveness of this approach, the team created a series of benchmarks that tested the models’ ability to perform well across different domains. The results showed that the composition of experts (CoE) outperformed individual LLMs in many cases, achieving better scores on tasks related to medicine, finance, and mathematics.


One of the key advantages of CoE is its ability to adapt to new tasks by combining models with complementary strengths. This allows it to learn from a broad range of data sources and generate more accurate responses.


The researchers also introduced a variant of CoE called Robust-CoE, which incorporates uncertainty quantification into the routing process. This helps the system identify when an expert’s response is uncertain or unclear, allowing it to request additional information or clarification.


While this approach shows promise, there are still challenges to overcome before it can be widely adopted. For example, determining the optimal composition of experts for a given task remains an open question.


Despite these limitations, the potential benefits of CoE and Robust-CoE make them worth exploring further. As AI systems continue to play a larger role in our lives, developing more efficient and effective ways to train and deploy them will be crucial for unlocking their full potential.


The researchers behind this work have made their code and datasets available online, allowing other developers to experiment with CoE and Robust-CoE on their own projects. This could lead to new breakthroughs and innovations in the field of natural language processing.


Ultimately, the composition of experts represents an exciting step forward in the development of AI systems that can learn from and adapt to complex tasks.


Cite this article: “Modularizing Language Models with Composition of Experts”, The Science Archive, 2025.


Language Models, Large Language Models, Composition Of Experts, Modular Ai, Natural Language Processing, Expert Systems, Machine Learning, Training Data, Uncertainty Quantification, Robust-Coe


Reference: Swayambhoo Jain, Ravi Raju, Bo Li, Zoltan Csaki, Jonathan Li, Kaizhao Liang, Guoyao Feng, Urmish Thakkar, Anand Sampat, Raghu Prabhakar, et al., “Composition of Experts: A Modular Compound AI System Leveraging Large Language Models” (2024).


Leave a Reply