MarkTechPost@AI 03月16日
SYMBOLIC-MOE: Mixture-of-Experts MoE Framework for Adaptive Instance-Level Mixing of Pre-Trained LLM Experts
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

SYMBOLIC-MOE是由UNC Chapel Hill的研究人员提出的一个基于符号、文本且无需梯度的混合专家框架,旨在实现预训练LLM专家的自适应实例级别混合。该框架强调细粒度的专业技能,例如数学中的代数或生物医学推理中的分子生物学。SYMBOLIC-MOE引入了一种基于技能的招聘策略,根据LLM专家已证明的优势,为每个特定的推理任务动态选择最相关的专家LLM。实验表明,SYMBOLIC-MOE在各种基准测试中均优于GPT4o-mini等强大的LLM以及多智能体方法,平均绝对改进幅度超过最佳多智能体基线8.15%。

🧠 SYMBOLIC-MOE框架包含模型画像创建与聚合器选择、专家招募和最终答案生成三个阶段,所有这些阶段都在推理过程中进行。

💡 SYMBOLIC-MOE引入了一种创新的批处理策略,以最大化吞吐量和效率。该策略首先分析所有实例以确定需要哪些LLM,然后根据所需专家智能地对问题实例进行分组,从而允许每个活跃的专家模型在单个批次中接收所有相关实例,并确保每个专家仅加载一次。

🏆 SYMBOLIC-MOE在各种基准测试中表现出色,始终优于所有基线方法,包括单模型策略、使用单模型的多智能体辩论以及MoA和ReConcile等多模型多智能体框架。它超过了最强的多智能体基线(Self-MoA),平均绝对改进幅度高达8.15%。

Like humans, large language models (LLMs) often have differing skills and strengths derived from differences in their architectures and training regimens. However, they struggle to combine specialized expertise across different domains, limiting their problem-solving capabilities compared to humans. Specialized models like MetaMath, WizardMath, and QwenMath excel at mathematical reasoning but often underperform on tasks requiring common sense or medical knowledge. Even within specific domains such as mathematics, models show nuanced variations in capability, e.g., one might excel at algebra while another masters geometry. creates a need for frameworks that can identify and select the most appropriate expert models for specific problems.

Existing approaches like Mixture-of-Experts (MoE) models distribute computation across multiple specialized components, with recent emphasis on sparse approaches that activate only the most relevant experts per input. The Sparse MoE (SMoE) method has improved efficiency across vision, language, and multimodal tasks but requires combining models in the parameter space through joint training. More recent frameworks like MoA (Mixture-of-Agents) attempt to address this by combining LLM outputs symbolically. Further, Multi-agent reasoning approaches have emerged as alternatives, such as the Student-teacher technique that distills reasoning capabilities from stronger to weaker agents, while debate frameworks allow multiple agents to refine arguments collectively.

Researchers from UNC Chapel Hill have proposed SYMBOLIC-MOE, a symbolic, text-based, and gradient-free Mixture-of-Experts framework to enable adaptive instance-level mixing of pre-trained LLM experts. It takes a fine-grained perspective by emphasizing specialized skills within broader domains like algebra within mathematics or molecular biology within biomedical reasoning. They also introduced a skill-based recruiting strategy that dynamically selects the most relevant expert LLMs for each specific reasoning task based on their demonstrated strengths. Moreover,  SYMBOLIC-MOE outperforms strong LLMs like GPT4o-mini, as well as multiagent approaches, with an absolute average improvement of 8.15% over the best multi-agent baseline.

SYMBOLIC-MOE consists of three stages: model profile creation and aggregator selection followed by expert recruitment and final answer generation, both of which take place during inference. To maximize throughput and efficiency, SYMBOLIC-MOE introduces an innovative batching strategy where all instances are first analyzed to determine which LLMs will be needed. The system then intelligently groups problem instances based on their required experts, allowing each active expert model to receive all relevant instances in a single batch and ensuring each expert is loaded only once. This solution enables efficient batched inference on a single GPU while supporting a diverse pool of 16 LLMs, with the flexibility to add more GPUs for further parallelization.

SYMBOLIC-MOE shows exceptional performance across diverse benchmarks. It consistently outperforms all baseline approaches, surpassing single-model strategies, multi-agent debates with a single model, and multi-model multi-agent frameworks like MoA and ReConcile. It exceeds the strongest multi-agent baseline (Self-MoA) by an impressive 8.15% absolute average improvement, 8.28% on MMLU-Pro, 13.45% on AIME, 4.92% on GPQA, and 6.08% on MedMCQA. SYMBOLIC-MOE achieves comparable or superior performance to larger models with 70B parameters by using four 7-8B parameter models. It outperforms Llama3.3 70B on AIME and GPQA while matching its performance on MedMCQA. Efficiency testing reveals that it operates 44% faster on a single GPU than MoA while achieving better accuracy.

In conclusion, researchers introduced SYMBOLIC-MOE, a scalable MoE framework that combines models through their symbolic output. This method identifies the skills needed for a given problem and recruits agents based on those skills to engage in a discussion about a given input. SYMBOLIC-MOE outperforms standard inference-time scaling methods as well as other debate frameworks and other mixture-of-agents methods, leading to strong performance across domains without human intervention. It’s average performance across heterogeneous tasks is in fact stronger than that of advanced proprietary models such as GPT4o-mini. However, this method has limitations: (a) It involves running multiple models, which increases inference cost, and (b) it relies on skills inferred from a small validation set to set the agent profiles.


    Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 80k+ ML SubReddit.

    The post SYMBOLIC-MOE: Mixture-of-Experts MoE Framework for Adaptive Instance-Level Mixing of Pre-Trained LLM Experts appeared first on MarkTechPost.

    Fish AI Reader

    Fish AI Reader

    AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

    FishAI

    FishAI

    鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

    联系邮箱 441953276@qq.com

    相关标签

    SYMBOLIC-MOE 混合专家 LLM 人工智能
    相关文章