LeanQuant commited on
Commit
24545db
·
verified ·
1 Parent(s): 4d84a84

Add files using upload-large-folder tool

Browse files
Files changed (1) hide show
  1. README.md +10 -0
README.md CHANGED
@@ -1,3 +1,13 @@
 
 
 
 
 
 
 
 
 
 
1
  ## DFloat11 Compressed Model: `Qwen/Qwen3-8B`
2
 
3
  This is a **losslessly compressed** version of [`Qwen/Qwen3-8B`](https://huggingface.co/Qwen/Qwen3-8B) using our custom **DFloat11** format. The outputs of this compressed model are **bit-for-bit identical** to the original BFloat16 model, while reducing GPU memory consumption by approximately **30%**.
 
1
+ ---
2
+ base_model: Qwen/Qwen3-8B
3
+ base_model_relation: quantized
4
+ tags:
5
+ - dfloat11
6
+ - df11
7
+ - lossless compression
8
+ - 70% size, 100% accuracy
9
+ ---
10
+
11
  ## DFloat11 Compressed Model: `Qwen/Qwen3-8B`
12
 
13
  This is a **losslessly compressed** version of [`Qwen/Qwen3-8B`](https://huggingface.co/Qwen/Qwen3-8B) using our custom **DFloat11** format. The outputs of this compressed model are **bit-for-bit identical** to the original BFloat16 model, while reducing GPU memory consumption by approximately **30%**.