merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

  - sources:
      - layer_range: [0, 10]
        model: meta-llama/Llama-3.2-1B-Instruct
        parameters:
          normalize: true
          int8_mask: true
          density: 0.5
          weight: 0.1
          random_seed: 0
          temperature: 0.5
          top_p: 0.65
          inference: true
          max_tokens: 999999999
          stream: true
        normalize: true
        int8_mask: true
        density: 0.5
        weight: 0.1
        random_seed: 0
        temperature: 0.5
        top_p: 0.65
        inference: true
        max_tokens: 999999999
        stream: true

parameters: 
  normalize: true
  int8_mask: true
  density: 0.5
  weight: 0.1
  random_seed: 0
  temperature: 0.5
  top_p: 0.65
  inference: true
  max_tokens: 999999999
  stream: true
normalize: true
int8_mask: true
density: 0.5
weight: 0.1
random_seed: 0
temperature: 0.5
top_p: 0.65
inference: true
max_tokens: 999999999
stream: true
merge_method: passthrough
dtype: bfloat16
Downloads last month
7
Safetensors
Model size
3.91B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Hjgugugjhuhjggg/mergekit-passthrough-trdtyuc

Finetuned
(867)
this model