ComfyUI_00159_

Velvet-Orchid-12B

This is a multi-stage merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Linear DELLA merge method using Vortex5/MegaMoon-Karcher-12B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

Merge Config (click to expand)
name: First  
models:
- model: Vortex5/Harmonic-Moon-12B
- model: Vortex5/Scarlet-Ink-12B
merge_method: karcher
dtype: bfloat16
parameters:
tol: 1e-10
max_iter: 1200
tokenizer:
source: union
base_model: Vortex5/MegaMoon-Karcher-12B
dtype: bfloat16
merge_method: della_linear
modules:
default:
  slices:
  - sources:
    - layer_range: [0, 40]
      model: ./intermediates/First
      parameters:
        density: 0.5
        epsilon: 0.4
        weight: [0.9, 0.2, 0.2, 0.2, 0.95, 0.8, 0.4, 0.2]
    - layer_range: [0, 40]
      model: taozi555/MN-12B-Mag-Mell-R1-KTO
      parameters:
        density: 0.5
        epsilon: 0.4
        weight: [0.1, 0.9, 1.0, 0.95, 0.4, 0.4, 0.5, 0.3]
    - layer_range: [0, 40]
      model: Vortex5/MegaMoon-Karcher-12B
parameters:
lambda: 1.0
normalize: 1.0
tokenizer: {}
Downloads last month
12
Safetensors
Model size
12B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Vortex5/Velvet-Orchid-12B