YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)

π Overview
A collection of optimized GGUF quantized models derived from demo, providing various performance-quality tradeoffs.
π Model Variants
Variant | Use Case | Download |
---|---|---|
demo_model_int8 | For mobile and embedded applications where memory and computational resources are limited, the int8 variant provides a good balance between accuracy and performance. | π₯ |
demo_model_int16 | For applications that require higher accuracy and can afford more computational resources, the int16 variant offers improved performance without significant memory overhead. | π₯ |
demo_model_fp16 | For high-performance computing applications where precision is crucial, the fp16 variant provides the best accuracy and is suitable for desktop and server environments. | π₯ |
π€ Contributors
Developed with β€οΈ by BlossomAI
Star βοΈ this repo if you find it valuable!
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support