Produced by Antigma Labs, Antigma Quantize Space
Follow Antigma Labs in X https://x.com/antigma_labs
Antigma's GitHub Homepage https://github.com/AntigmaLabs
llama.cpp quantization
Using llama.cpp release b4944 for quantization. Original model: https://huggingface.co/Qwen/Qwen3-14B Run them directly with llama.cpp, or any other llama.cpp based project
Prompt format
<๏ฝbeginโofโsentence๏ฝ>{system_prompt}<๏ฝUser๏ฝ>{prompt}<๏ฝAssistant๏ฝ><๏ฝendโofโsentence๏ฝ><๏ฝAssistant๏ฝ>
Download a file (not the whole branch) from below:
Filename | Quant type | File Size | Split |
---|---|---|---|
qwen3-14b-q4_k_m.gguf | Q4_K_M | 8.38 GB | False |
qwen3-14b-q4_0.gguf | Q4_0 | 7.93 GB | False |
qwen3-14b-q4_k_s.gguf | Q4_K_S | 7.98 GB | False |
Downloading using huggingface-cli
Click to view download instructions
First, make sure you have hugginface-cli installed:pip install -U "huggingface_hub[cli]"
Then, you can target the specific file you want:
huggingface-cli download https://huggingface.co/Antigma/Qwen3-14B-GGUF --include "qwen3-14b-q4_k_m.gguf" --local-dir ./
If the model is bigger than 50GB, it will have been split into multiple files. In order to download them all to a local folder, run:
huggingface-cli download https://huggingface.co/Antigma/Qwen3-14B-GGUF --include "qwen3-14b-q4_k_m.gguf/*" --local-dir ./
You can either specify a new local-dir (deepseek-ai_DeepSeek-V3-0324-Q8_0) or download them all in place (./)
- Downloads last month
- 16
Hardware compatibility
Log In
to view the estimation
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support