Grafted-Wind-Elementals-2x70B

A MoE merge of:

Why use one 70B when you could use two?
I really liked this model's writing ability. Just need a few more 3090s, right?

License: Non-commercial research use only.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 76.21
AI2 Reasoning Challenge (25-Shot) 73.38
HellaSwag (10-Shot) 89.08
MMLU (5-Shot) 75.79
TruthfulQA (0-shot) 65.57
Winogrande (5-shot) 84.85
GSM8k (5-shot) 68.61
Downloads last month
4
Safetensors
Model size
125B params
Tensor type
FP16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for lodrick-the-lafted/Grafted-Wind-Elementals-2x70B

Finetuned
(25)
this model
Quantizations
2 models

Evaluation results