Image by Author
# Introduction
Merging language models is one of the most powerful techniques for improving AI performance without costly retraining. By combining two or more pre-trained models, you can create a single model that inherits the best capabilities from each parent. In this tutorial, you will learn how to merge large language models (LLMs) easily using Unsloth Studio, a free, no-code web interface that runs entirely on your computer.
# Defining Unsloth Studio
Unsloth Studio is an open-source, browser-based graphical user interface (GUI) launched in March 2026 by Unsloth AI. It allows you to run, fine-tune, and export LLMs without writing a single line of code. Here is what makes it special:
- No coding required — all operations happen through a visual interface
- Runs 100% locally — your data never leaves your computer
- Fast and memory-efficient — up to 2x faster training with 70% less video random access memory (VRAM) usage compared to traditional methods
- Cross-platform — works on Windows, Linux, macOS, and Windows Subsystem for Linux (WSL)
Unsloth Studio supports popular models including Llama, Qwen, Gemma, DeepSeek, Mistral, and hundreds more.
# Understanding Why Language Models Are Merged
Before exploring the Unsloth Studio tutorial, it is important to understand why model merging matters.
When you fine-tune a model for a specific task (e.g. coding, customer service, or medical Q&A), you create low-rank adaptation (LoRA) adapters that change the original model’s behavior. The challenge is that you might have multiple adapters, each working well at different tasks. How do you combine them into one powerful model?
Model merging solves this problem. Instead of juggling multiple adapters, merging combines their capabilities into a single, deployable model. Here are common use cases:
- Combine a math-specialized model with a code-specialized model to create a model that excels at both
- Merge a model fine-tuned on English data with one fine-tuned on multilingual data
- Blend a creative writing model with a factual Q&A model
According to NVIDIA’s technical blog on model merging, merging combines the weights of multiple customized LLMs, increasing resource utilization and adding value to successful models.
// Prerequisites
Before starting, ensure your system meets the following requirements:
- NVIDIA graphics processing unit (GPU) (RTX 30, 40, or 50 series recommended) for training, though central processing unit (CPU)-only works for basic inference
- Python 3.10+ with pip and at least 16GB of random access memory (RAM)
- 20–50GB of free storage space (depending on the model size); and the models themselves, either one base model plus one or more fine-tuned LoRA adapters, or multiple pre-trained models you wish to merge.
# Getting Started with Unsloth Studio
Setting up Unsloth Studio is straightforward. Use a dedicated Conda environment to avoid dependency conflicts. Run conda create -n unsloth_env python=3.10 followed by conda activate unsloth_env before installation.
// Installing via pip
Open your terminal and run:
For Windows, ensure you have PyTorch installed first. The official Unsloth documentation provides detailed platform-specific instructions.
// Launching Unsloth Studio
After installation, start the Studio with:
The first run compiles llama.cpp binaries, which takes about 5–10 minutes. Once complete, a browser window opens automatically with the Unsloth Studio dashboard.
// Verifying the Installation
To confirm everything works, run:
You should see a welcome message with version information. For example, Unsloth version 2025.4.1 running on Compute Unified Device Architecture (CUDA) with optimized kernels.
# Exploring Model Merging Techniques
Unsloth Studio supports three main merging methods. Each has unique strengths, and choosing the right one depends on your goals.
// SLERP (Spherical Linear Interpolation)
SLERP is best for merging exactly two models with smooth, balanced results. SLERP performs interpolation along a geodesic path in weight space, preserving geometric properties better than simple averaging. Think of it as a “smooth blend” between two models.
Key characteristics:
- Only merges two models at a time
- Preserves the unique characteristics of both parents
- Great for combining models from the same family (e.g. Mistral v0.1 with Mistral v0.2)
// TIES-Merging (Trim, Elect Sign, and Merge)
TIES-Merging is for merging three or more models while resolving conflicts. TIES-Merging was introduced to address two major problems in model merging:
- Redundant parameter values that waste capacity
- Disagreements on the sign (positive/negative direction) of parameters across models
The method works in three steps:
- Trim — keep only parameters that changed significantly during fine-tuning
- Elect Sign — determine the majority direction for each parameter across models
- Merge — combine only parameters that align with the agreed sign
Research shows TIES-Merging as the most effective and robust method among available techniques.
// DARE (Drop And REscale)
This is also best for merging models that have many redundant parameters. DARE randomly drops a percentage of delta parameters and rescales the remaining ones. This reduces interference and often improves performance, especially when merging multiple models. DARE is typically used as a pre-processing step before TIES (creating DARE-TIES).
NOTE: Language models have extreme redundancy; DARE can eliminate 90% or even 99% of delta parameters without significant performance loss.
// Comparing Merging Methods
Method
Best For
Number of Models
Key Advantage
SLERP
Two similar models
Exactly 2
Smooth, balanced blend
TIES
3+ models, task-specific
Multiple
Resolves sign conflicts
DARE
Redundant parameters
Multiple
Reduces interference
# Merging Models in Unsloth Studio
Now for the practical part of model merging. Follow these steps to perform your first merge.
// Launching Unsloth Studio and Navigating to Training
Open your browser and go to http://localhost:3000 (or the address shown after launching). Click on the Training module from the dashboard.
// Selecting or Creating a Training Run
In Unsloth Studio, a training run represents a complete training session that may contain multiple checkpoints. To merge:
- If you already have a training run with LoRA adapters, select it from the list
- If you’re starting fresh, create a new run and load your base model
Each run contains checkpoints — saved versions of your model at different training stages. Later checkpoints typically represent the final trained model, but you can select any checkpoint for merging.
// Choosing the Merge Method
Navigate to the Export section of the Studio. Here you’ll see three export types:
- Merged Model — 16-bit model with LoRA adapter merged into base weights
- LoRA Only — exports only adapter weights (requires original base model)
- GGUF — converts to GGUF format for llama.cpp or Ollama inference
For model merging, select Merged Model.
As of the latest documentation, Unsloth Studio primarily supports merging LoRA adapters into base models. For advanced techniques like SLERP or TIES merging of multiple full models, you may need to use MergeKit alongside Unsloth. Many developers fine-tune multiple LoRAs with Unsloth, then use MergeKit for SLERP or TIES merging.
// Configuring Low-Rank Adaptation Merge Settings
Depending on the chosen method, different options will appear. For LoRA merging (the simplest method):
- Select the LoRA adapter to merge
- Choose output precision (16-bit or 4-bit)
- Set save location
For advanced merging with MergeKit (if using the command-line interface (CLI)):
- Define the base model path
- List parent models to merge
- Set merge method (SLERP, TIES, or DARE)
- Configure interpolation parameters
Here’s an example of what a MergeKit configuration looks like (for reference):
merge_method: ties
base_model: path/to/base/model
models:
– model: path/to/model1
parameters:
weight: 1.0
– model: path/to/model2
parameters:
weight: 0.5
dtype: bfloat16
// Executing the Merge
Click Export or Merge to start the process. Unsloth Studio merges LoRA weights using the formula:
\[
W_{\text{merged}} = W_{\text{base}} + (A \cdot B) \times \text{scaling}
\]
Where:
- \( W_{\text{base}} \) is the original weight matrix
- \( A \) and \( B \) are LoRA adapter matrices
- Scaling is the LoRA scaling factor (typically lora_alpha / lora_r)
For 4-bit models, Unsloth dequantizes to FP32, performs the merge, and then requantizes back to 4-bit — all automatically.
// Saving and Exporting the Merged Model
Once the merging is complete, two options are available:
- Save Locally — downloads the merged model files to your machine for local deployment
- Push to Hub — uploads directly to Hugging Face Hub for sharing and collaboration (requires a Hugging Face write token)
The merged model is saved in safetensors format by default, compatible with llama.cpp, vLLM, Ollama, and LM Studio.
# Best Practices for Successful Model Merging
Based on community experience and research findings, here are proven tips:
- Start with Compatible Models
Models from the same architecture family (e.g. both based on Llama) merge more successfully than cross-architecture merges - Use DARE as a Pre-processor
When merging multiple models, apply DARE first to eliminate redundant parameters, then TIES for final merging. This DARE-TIES combination is widely used in the community - Experiment with Interpolation Parameters
For SLERP merges, the interpolation factor \( t \) determines the blend:- \( t = 0 \rightarrow \) Model A only
- \( t = 0.5 \rightarrow \) Equal blend
- \( t = 1 \rightarrow \) Model B only
Start with \( t = 0.5 \) and adjust based on your needs
- Evaluate Before Deploying
Always test your merged model against a benchmark. Unsloth Studio includes a Model Arena that lets you compare two models side-by-side with the same prompt - Watch Your Disk Space
Merging large models (like 70B parameters) can temporarily require significant disk space. The merge process creates intermediate files that may require up to 2–3x the model’s size temporarily
# Conclusion
In this article, you have learned that merging language models with Unsloth Studio opens up powerful possibilities for AI practitioners. You can now combine the strengths of multiple specialized models into one efficient, deployable model — all without writing complex code.
To recap what was covered:
- Unsloth Studio is a no-code, local web interface for AI model training and merging
- Merging models allows you to combine capabilities from multiple adapters without retraining
- Three key techniques include SLERP (smooth blend of two models), TIES (resolve conflicts across many), and DARE (reduce redundancy)
- The merge process is a clear 6-step process from installation to export
Download Unsloth Studio and try combining your first two models today.
Shittu Olumide is a software engineer and technical writer passionate about leveraging cutting-edge technologies to craft compelling narratives, with a keen eye for detail and a knack for simplifying complex concepts. You can also find Shittu on Twitter.

