LeanQuant commited on
Commit
ff6c02c
·
verified ·
1 Parent(s): 68dce0d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -2
README.md CHANGED
@@ -1,7 +1,13 @@
1
  ---
2
- base_model:
3
- - google/gemma-3-4b-it
 
 
 
 
 
4
  ---
 
5
  ## DFloat11 Compressed Model: `google/gemma-3-4b-it`
6
 
7
  This is a **losslessly compressed** version of [`google/gemma-3-4b-it`](https://huggingface.co/google/gemma-3-4b-it) using our custom **DFloat11** format. The outputs of this compressed model are **bit-for-bit identical** to the original BFloat16 model, while reducing GPU memory consumption by approximately **30%**.
 
1
  ---
2
+ base_model: google/gemma-3-4b-pt
3
+ base_model_relation: quantized
4
+ tags:
5
+ - dfloat11
6
+ - df11
7
+ - lossless compression
8
+ - 70% size, 100% accuracy
9
  ---
10
+
11
  ## DFloat11 Compressed Model: `google/gemma-3-4b-it`
12
 
13
  This is a **losslessly compressed** version of [`google/gemma-3-4b-it`](https://huggingface.co/google/gemma-3-4b-it) using our custom **DFloat11** format. The outputs of this compressed model are **bit-for-bit identical** to the original BFloat16 model, while reducing GPU memory consumption by approximately **30%**.