merge

It sucks. Alas.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using mergekit-community/L3.3-Oddbase-70B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
 - model: mergekit-community/ignore_L3.x-Monk-70B
   parameters:
     density: 0.7
     weight: 0.5
 - model: mergekit-community/ignore_L3.x-Hero-70B
   parameters:
     density: 0.9
     weight: 1
 - model: mergekit-community/ignore_L3.1-Deity-70B
   parameters:
     density: 0.5
     weight: 0.7
merge_method: dare_ties
base_model: mergekit-community/L3.3-Oddbase-70B
tokenizer_source: base
Downloads last month
16
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Yobenboben/L3.3-70B-Meg-Mall