Skip to content

Model Merging

PMetal supports 16 model merging strategies (12 via CLI, 4 library-only). Features GPU-accelerated merging, FP8-aware operations, and async double-buffered streaming for large models.

MethodDescription
linearSimple weighted averaging
slerpSpherical linear interpolation
tiesTask arithmetic with sparsification and sign consensus
dare_tiesRandom pruning with rescaling (TIES variant)
dare_linearRandom pruning with rescaling (linear variant)
task_arithmeticTask vector arithmetic
dellaAdaptive magnitude-based pruning
della_linearAdaptive magnitude pruning (linear variant)
breadcrumbsBreadcrumbs merge strategy
model_stockGeometric interpolation based on task vector similarity
nearswapNear-swap merge strategy
passthroughLayer passthrough composition
StrategyDescription
RamMergeRAM merge strategy
SouperMergeSouper merge strategy
KarcherMergeKarcher mean on weight manifold
MultiSlerpMergeMulti-model SLERP
Terminal window
# SLERP merge of two models
pmetal merge --models model-a model-b --method slerp --t 0.5
# TIES with sparsification
pmetal merge --models base ft-1 ft-2 --method ties --density 0.5
# DARE-TIES
pmetal merge --models model-a model-b --method dare_ties --density 0.7
  • GPU-Accelerated Merging: Metal-based merge operations for large models
  • FP8-Aware Merging: Merge with FP8 quantization for memory efficiency
  • Async Merge Pipeline: Double-buffered streaming for models that exceed memory
  • LoRA Fusing: Merge LoRA adapters into base weights (standard and accurate modes)