Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -20,9 +20,9 @@ A collection of optimized GGUF quantized models derived from [bloomvn-0.5b-ppo](
|
|
20 |
|
21 |
| Variant | Use Case | Download |
|
22 |
|---------|-----------|------------|
|
23 |
-
| base | This
|
24 |
-
| q2_k | The 2-bit quantized
|
25 |
-
| q3_k_m | The 3-bit quantized
|
26 |
|
27 |
## π€ Contributors
|
28 |
|
|
|
20 |
|
21 |
| Variant | Use Case | Download |
|
22 |
|---------|-----------|------------|
|
23 |
+
| base | This FP16 format model is ideal for applications where high accuracy is crucial and storage space is not a concern, such as in data centers or high-performance computing environments. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/base.gguf)
|
24 |
+
| q2_k | The 2-bit quantized model is suitable for extremely constrained environments, such as low-power edge devices or those with very limited memory, where a balance between size and performance is necessary. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q2_k.gguf)
|
25 |
+
| q3_k_m | The 3-bit quantized model offers a good balance between compression and performance, making it suitable for memory-limited devices that still require a reasonable level of accuracy, such as mid-range smartphones or embedded systems. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q3_k_m.gguf)
|
26 |
|
27 |
## π€ Contributors
|
28 |
|