Qwen3-1.7B-HORROR-Imatrix-Max-GGUF

HORROR Imatrix Quants of new "Qwen 3 - 1.7B" model with MAX "output tensor" at BF16 to improve reasoning / output generation.

HORROR Imatrix dataset was generated in house. It was generated using "Grand Horror 16B" model.

Imatrix effect will be stronger, the lower the quant you use with IQ4XS/IQ4NL being the best balanced quant for quality and Imatrix effect.

These quants will also be the strongest for creative use cases.

For stronger reasoning use higher quants.

Q8_0 quant is maxed only, as Imatrix has no effect on this quant.

F16 is full precision.

Context Length: 32 K + 8K output generation.

NOTE - Jinja Template / Template to Use with this Model:

If you are having issues with Jinja "auto template", use CHATML template.

OR (LMSTUDIO users / option)

Update the Jinja Template (go to this site, template-> copy the "Jinja template" and then paste.)

[ https://lmstudio.ai/neil/qwen3-thinking ]

Other Notes:

Reasoning is ON by default in this model, and model will auto-generate "think" block(s).

For benchmarks, usage info, settings please see org model card here:

[ https://huggingface.co/Qwen/Qwen3-1.7B ]

[ Model card, and examples to follow. ]

Downloads last month
34
GGUF
Model size
2.03B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DavidAU/Qwen3-1.7B-HORROR-Imatrix-Max-GGUF

Finetuned
Qwen/Qwen3-1.7B
Quantized
(46)
this model

Collections including DavidAU/Qwen3-1.7B-HORROR-Imatrix-Max-GGUF