Vuanhngo11 commited on
Commit
de59f41
Β·
verified Β·
1 Parent(s): ec24bd7

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -20,9 +20,9 @@ A collection of optimized GGUF quantized models derived from [bloomvn-0.5b-ppo](
20
 
21
  | Variant | Use Case | Download |
22
  |---------|-----------|------------|
23
- | base | This FP16 format model is ideal for applications where high accuracy is crucial and storage space is not a concern, such as in data centers or high-performance computing environments. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/base.gguf)
24
- | q2_k | The 2-bit quantized model is suitable for extremely constrained environments, such as low-power edge devices or those with very limited memory, where a balance between size and performance is necessary. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q2_k.gguf)
25
- | q3_k_m | The 3-bit quantized model offers a good balance between compression and performance, making it suitable for memory-limited devices that still require a reasonable level of accuracy, such as mid-range smartphones or embedded systems. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q3_k_m.gguf)
26
 
27
  ## 🀝 Contributors
28
 
 
20
 
21
  | Variant | Use Case | Download |
22
  |---------|-----------|------------|
23
+ | base | The base model is suitable for applications where model size is not a concern, and high accuracy is required. It can be used for tasks such as text generation, language translation, and text summarization. This model is in FP16 format, providing a good balance between size and performance. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/base.gguf)
24
+ | q2_k | The q2_k variant is ideal for extremely constrained environments where model size is a significant concern, such as embedded systems or low-end mobile devices. Although it is highly compressed, it still maintains a reasonable level of accuracy, making it suitable for simple language tasks. This 2-bit quantized model is a good choice when storage space is limited. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q2_k.gguf)
25
+ | q3_k_m | The q3_k_m variant is designed for memory-limited devices that require a balance between model size and accuracy. This 3-bit quantized model is very compressed, making it suitable for mid-range mobile devices or systems with limited storage capacity. It is a good choice for applications where a moderate level of accuracy is required, such as language understanding or text classification. | [πŸ“₯](https://huggingface.co/Vuanhngo11/bloomvn-0.5b-ppo-gguf/resolve/main/q3_k_m.gguf)
26
 
27
  ## 🀝 Contributors
28