Merging Language Models with Unsloth Studio

Merge LLMs easily with Unsloth Studio's no-code GUI and combine models without retraining.



Merging Language Models with Unsloth Studio
Image by Author

 

Introduction

 
Merging language models is one of the most powerful techniques for improving AI performance without costly retraining. By combining two or more pre-trained models, you can create a single model that inherits the best capabilities from each parent. In this tutorial, you will learn how to merge large language models (LLMs) easily using Unsloth Studio, a free, no-code web interface that runs entirely on your computer.

 

Defining Unsloth Studio

 
Unsloth Studio is an open-source, browser-based graphical user interface (GUI) launched in March 2026 by Unsloth AI. It allows you to run, fine-tune, and export LLMs without writing a single line of code. Here is what makes it special:

  • No coding required — all operations happen through a visual interface
  • Runs 100% locally — your data never leaves your computer
  • Fast and memory-efficient — up to 2x faster training with 70% less video random access memory (VRAM) usage compared to traditional methods
  • Cross-platform — works on Windows, Linux, macOS, and Windows Subsystem for Linux (WSL)

Unsloth Studio supports popular models including Llama, Qwen, Gemma, DeepSeek, Mistral, and hundreds more.

 

Understanding Why Language Models Are Merged

 
Before exploring the Unsloth Studio tutorial, it is important to understand why model merging matters.
When you fine-tune a model for a specific task (e.g. coding, customer service, or medical Q&A), you create low-rank adaptation (LoRA) adapters that change the original model's behavior. The challenge is that you might have multiple adapters, each working well at different tasks. How do you combine them into one powerful model?

Model merging solves this problem. Instead of juggling multiple adapters, merging combines their capabilities into a single, deployable model. Here are common use cases:

  • Combine a math-specialized model with a code-specialized model to create a model that excels at both
  • Merge a model fine-tuned on English data with one fine-tuned on multilingual data
  • Blend a creative writing model with a factual Q&A model

According to NVIDIA's technical blog on model merging, merging combines the weights of multiple customized LLMs, increasing resource utilization and adding value to successful models.

 

// Prerequisites

Before starting, ensure your system meets the following requirements:

  • NVIDIA graphics processing unit (GPU) (RTX 30, 40, or 50 series recommended) for training, though central processing unit (CPU)-only works for basic inference
  • Python 3.10+ with pip and at least 16GB of random access memory (RAM)
  • 20–50GB of free storage space (depending on the model size); and the models themselves, either one base model plus one or more fine-tuned LoRA adapters, or multiple pre-trained models you wish to merge.

 

Getting Started with Unsloth Studio

 
Setting up Unsloth Studio is straightforward. Use a dedicated Conda environment to avoid dependency conflicts. Run conda create -n unsloth_env python=3.10 followed by conda activate unsloth_env before installation.

 

// Installing via pip

Open your terminal and run:

pip install unsloth

 
For Windows, ensure you have PyTorch installed first. The official Unsloth documentation provides detailed platform-specific instructions.

 

// Launching Unsloth Studio

After installation, start the Studio with:

unsloth studio setup

 

The first run compiles llama.cpp binaries, which takes about 5–10 minutes. Once complete, a browser window opens automatically with the Unsloth Studio dashboard.

 

// Verifying the Installation

To confirm everything works, run:

python -m unsloth

 

You should see a welcome message with version information. For example, Unsloth version 2025.4.1 running on Compute Unified Device Architecture (CUDA) with optimized kernels.

 

Exploring Model Merging Techniques

 
Unsloth Studio supports three main merging methods. Each has unique strengths, and choosing the right one depends on your goals.

 

// SLERP (Spherical Linear Interpolation)

SLERP is best for merging exactly two models with smooth, balanced results. SLERP performs interpolation along a geodesic path in weight space, preserving geometric properties better than simple averaging. Think of it as a "smooth blend" between two models.

Key characteristics:

  • Only merges two models at a time
  • Preserves the unique characteristics of both parents
  • Great for combining models from the same family (e.g. Mistral v0.1 with Mistral v0.2)

 

// TIES-Merging (Trim, Elect Sign, and Merge)

TIES-Merging is for merging three or more models while resolving conflicts. TIES-Merging was introduced to address two major problems in model merging:

  • Redundant parameter values that waste capacity
  • Disagreements on the sign (positive/negative direction) of parameters across models

The method works in three steps:

  • Trim — keep only parameters that changed significantly during fine-tuning
  • Elect Sign — determine the majority direction for each parameter across models
  • Merge — combine only parameters that align with the agreed sign

Research shows TIES-Merging as the most effective and robust method among available techniques.

 

// DARE (Drop And REscale)

This is also best for merging models that have many redundant parameters. DARE randomly drops a percentage of delta parameters and rescales the remaining ones. This reduces interference and often improves performance, especially when merging multiple models. DARE is typically used as a pre-processing step before TIES (creating DARE-TIES).

NOTE: Language models have extreme redundancy; DARE can eliminate 90% or even 99% of delta parameters without significant performance loss.

 

// Comparing Merging Methods

Method Best For Number of Models Key Advantage
SLERP Two similar models Exactly 2 Smooth, balanced blend
TIES 3+ models, task-specific Multiple Resolves sign conflicts
DARE Redundant parameters Multiple Reduces interference

 

 

Merging Models in Unsloth Studio

 
Now for the practical part of model merging. Follow these steps to perform your first merge.

 

// Launching Unsloth Studio and Navigating to Training

Open your browser and go to http://localhost:3000 (or the address shown after launching). Click on the Training module from the dashboard.

 

// Selecting or Creating a Training Run

In Unsloth Studio, a training run represents a complete training session that may contain multiple checkpoints. To merge:

  • If you already have a training run with LoRA adapters, select it from the list
  • If you're starting fresh, create a new run and load your base model

Each run contains checkpoints — saved versions of your model at different training stages. Later checkpoints typically represent the final trained model, but you can select any checkpoint for merging.

 

// Choosing the Merge Method

Navigate to the Export section of the Studio. Here you'll see three export types:

  • Merged Model — 16-bit model with LoRA adapter merged into base weights
  • LoRA Only — exports only adapter weights (requires original base model)
  • GGUF — converts to GGUF format for llama.cpp or Ollama inference

For model merging, select Merged Model.

As of the latest documentation, Unsloth Studio primarily supports merging LoRA adapters into base models. For advanced techniques like SLERP or TIES merging of multiple full models, you may need to use MergeKit alongside Unsloth. Many developers fine-tune multiple LoRAs with Unsloth, then use MergeKit for SLERP or TIES merging.

 

// Configuring Low-Rank Adaptation Merge Settings

Depending on the chosen method, different options will appear. For LoRA merging (the simplest method):

  • Select the LoRA adapter to merge
  • Choose output precision (16-bit or 4-bit)
  • Set save location

For advanced merging with MergeKit (if using the command-line interface (CLI)):

  • Define the base model path
  • List parent models to merge
  • Set merge method (SLERP, TIES, or DARE)
  • Configure interpolation parameters

Here's an example of what a MergeKit configuration looks like (for reference):

merge_method: ties
base_model: path/to/base/model
models:
  - model: path/to/model1
    parameters:
      weight: 1.0
  - model: path/to/model2
    parameters:
      weight: 0.5
dtype: bfloat16

 

 

// Executing the Merge

Click Export or Merge to start the process. Unsloth Studio merges LoRA weights using the formula:

\[
W_{\text{merged}} = W_{\text{base}} + (A \cdot B) \times \text{scaling}
\]

Where:

  • \( W_{\text{base}} \) is the original weight matrix
  • \( A \) and \( B \) are LoRA adapter matrices
  • Scaling is the LoRA scaling factor (typically lora_alpha / lora_r)

For 4-bit models, Unsloth dequantizes to FP32, performs the merge, and then requantizes back to 4-bit — all automatically.

 

// Saving and Exporting the Merged Model

Once the merging is complete, two options are available:

  • Save Locally — downloads the merged model files to your machine for local deployment
  • Push to Hub — uploads directly to Hugging Face Hub for sharing and collaboration (requires a Hugging Face write token)

The merged model is saved in safetensors format by default, compatible with llama.cpp, vLLM, Ollama, and LM Studio.

 

Best Practices for Successful Model Merging

 
Based on community experience and research findings, here are proven tips:

  1. Start with Compatible Models
    Models from the same architecture family (e.g. both based on Llama) merge more successfully than cross-architecture merges
  2. Use DARE as a Pre-processor
    When merging multiple models, apply DARE first to eliminate redundant parameters, then TIES for final merging. This DARE-TIES combination is widely used in the community
  3. Experiment with Interpolation Parameters
    For SLERP merges, the interpolation factor \( t \) determines the blend:

    • \( t = 0 \rightarrow \) Model A only
    • \( t = 0.5 \rightarrow \) Equal blend
    • \( t = 1 \rightarrow \) Model B only

    Start with \( t = 0.5 \) and adjust based on your needs

  4. Evaluate Before Deploying
    Always test your merged model against a benchmark. Unsloth Studio includes a Model Arena that lets you compare two models side-by-side with the same prompt
  5. Watch Your Disk Space
    Merging large models (like 70B parameters) can temporarily require significant disk space. The merge process creates intermediate files that may require up to 2–3x the model's size temporarily

 

Conclusion

 
In this article, you have learned that merging language models with Unsloth Studio opens up powerful possibilities for AI practitioners. You can now combine the strengths of multiple specialized models into one efficient, deployable model — all without writing complex code.

To recap what was covered:

  • Unsloth Studio is a no-code, local web interface for AI model training and merging
  • Merging models allows you to combine capabilities from multiple adapters without retraining
  • Three key techniques include SLERP (smooth blend of two models), TIES (resolve conflicts across many), and DARE (reduce redundancy)
  • The merge process is a clear 6-step process from installation to export

Download Unsloth Studio and try combining your first two models today.
 
 

Shittu Olumide is a software engineer and technical writer passionate about leveraging cutting-edge technologies to craft compelling narratives, with a keen eye for detail and a knack for simplifying complex concepts. You can also find Shittu on Twitter.


Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy


Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy

Get the FREE ebook 'KDnuggets Artificial Intelligence Pocket Dictionary' along with the leading newsletter on Data Science, Machine Learning, AI & Analytics straight to your inbox.

By subscribing you accept KDnuggets Privacy Policy

No, thanks!