Velvet-Orchid-12B
This is a multi-stage merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the Linear DELLA merge method using Vortex5/MegaMoon-Karcher-12B as a base.
Models Merged
The following models were included in the merge:
- ./intermediates/First
- taozi555/MN-12B-Mag-Mell-R1-KTO
- Vortex5/Scarlet-Ink-12B
- Vortex5/Harmonic-Moon-12B
Configuration
The following YAML configuration was used to produce this model:
Merge Config (click to expand)
name: First
models:
- model: Vortex5/Harmonic-Moon-12B
- model: Vortex5/Scarlet-Ink-12B
merge_method: karcher
dtype: bfloat16
parameters:
tol: 1e-10
max_iter: 1200
tokenizer:
source: union
base_model: Vortex5/MegaMoon-Karcher-12B
dtype: bfloat16
merge_method: della_linear
modules:
default:
slices:
- sources:
- layer_range: [0, 40]
model: ./intermediates/First
parameters:
density: 0.5
epsilon: 0.4
weight: [0.9, 0.2, 0.2, 0.2, 0.95, 0.8, 0.4, 0.2]
- layer_range: [0, 40]
model: taozi555/MN-12B-Mag-Mell-R1-KTO
parameters:
density: 0.5
epsilon: 0.4
weight: [0.1, 0.9, 1.0, 0.95, 0.4, 0.4, 0.5, 0.3]
- layer_range: [0, 40]
model: Vortex5/MegaMoon-Karcher-12B
parameters:
lambda: 1.0
normalize: 1.0
tokenizer: {}
- Downloads last month
- 12
Model tree for Vortex5/Velvet-Orchid-12B
Merge model
this model
