Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -20,9 +20,9 @@ A collection of optimized GGUF quantized models derived from [bloomvn-0.5b-ppo](
|
|
20 |
|
21 |
| Variant | Use Case | Download |
|
22 |
|---------|-----------|------------|
|
23 |
-
| base |
|
24 |
-
| q2_k | The
|
25 |
-
| q3_k_m | The
|
26 |
|
27 |
## π€ Contributors
|
28 |
|
|
|
20 |
|
21 |
| Variant | Use Case | Download |
|
22 |
|---------|-----------|------------|
|
23 |
+
| base | The base model is suitable for applications where model size is not a concern, and high accuracy is required. It can be used for tasks such as text generation, language translation, and text summarization. This model is in FP16 format, providing a good balance between size and performance. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/base.gguf)
|
24 |
+
| q2_k | The q2_k variant is ideal for extremely constrained environments where model size is a significant concern, such as embedded systems or low-end mobile devices. Although it is highly compressed, it still maintains a reasonable level of accuracy, making it suitable for simple language tasks. This 2-bit quantized model is a good choice when storage space is limited. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q2_k.gguf)
|
25 |
+
| q3_k_m | The q3_k_m variant is designed for memory-limited devices that require a balance between model size and accuracy. This 3-bit quantized model is very compressed, making it suitable for mid-range mobile devices or systems with limited storage capacity. It is a good choice for applications where a moderate level of accuracy is required, such as language understanding or text classification. | [π₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q3_k_m.gguf)
|
26 |
|
27 |
## π€ Contributors
|
28 |
|