Skip to main content
    Skip to main contentSkip to navigationSkip to footer
    Artificial Intelligence

    Model Merging

    Also known as:
    Model Fusion
    Weight Averaging
    Model Soup
    Model Combination
    Updated: 2/11/2026

    Techniques for combining multiple trained models into a single model that unifies the strengths of all source models – without additional training.

    Quick Summary

    Model merging combines multiple trained models into one – stack capabilities without extra training through weight averaging, SLERP, or task arithmetic.

    Explanation

    Model merging averages weights of multiple models (linear, SLERP, TIES, DARE). "Model Soup" combines fine-tuning checkpoints. Task arithmetic adds/subtracts task vectors. Enables capability stacking without compute explosion.

    Marketing Relevance

    Hot trend in open-source LLM community: Merged models dominate leaderboards. Marketing teams can combine specialized models (coding, creativity, German) into custom assistants.

    Example

    A team merges a German language model with a creative writing model and a fact-focused model. The result: A marketing assistant that generates creative German texts with high factual accuracy.

    Common Pitfalls

    Only works with models of the same architecture. Not all capabilities transfer cleanly. Can lead to interference between tasks. Quality of merge method is critical.

    Origin & History

    Wortsman et al. (2022) coined "Model Soups" for averaged fine-tuning checkpoints. Ilharco et al. (2022) introduced task arithmetic. TIES-Merging (Yadav et al., 2023) and DARE (Yu et al., 2023) improved merge quality. In 2024, merged models dominate open-source leaderboards.

    Comparisons & Differences

    Model Merging vs. Ensemble Learning

    Ensembles run multiple models in parallel (N× cost); merging creates a single model (1× cost) from multiple.

    Model Merging vs. Knowledge Distillation

    Distillation trains a new model from a teacher; merging combines weights without additional training.

    Related Services

    Related Terms

    👋Questions? Chat with us!