-
-
-
-
-
-
Inference Providers
Active filters:
ollama
prithivMLmods/Llama-3.2-1B-GGUF
Text Generation
•
Updated
•
85
•
2
DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters
Updated
•
110
itlwas/Llama-SmolTalk-3.2-1B-Instruct-Q4_K_M-GGUF
Text Generation
•
Updated
•
19
•
1
andresdegante/papalia3
Text Generation
•
Updated
•
1
GainEnergy/ogai-reasoner
Text Generation
•
Updated
•
3
mradermacher/ReasonableLlama3-3B-Jr-GGUF
dipeshmajithia/mirror_dolly
Updated
•
7
•
1
DevItachi/Robin
Updated
•
121
•
2
pacozaa/mistral-unsloth-chatml-first
Updated
•
47
pacozaa/tinyllama-alpaca-lora
Updated
pacozaa/bonito-gguf
pacozaa/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF
Updated
•
46
pacozaa/mistral-sharegpt90k
pacozaa/mistral-sharegpt90k-merged_16bit
Text Generation
•
Updated
•
2
TrabEsrever/dolphin-2.9-llama3-70b-GGUF
Updated
daekeun-ml/Phi-3-medium-4k-instruct-ko-poc-gguf-v0.1
Text Generation
•
Updated
•
15
•
1
hierholzer/Llama-3.1-70B-Instruct-GGUF
Text Generation
•
Updated
•
50
•
3
LucasInsight/Meta-Llama-3.1-8B-Instruct
Updated
•
5
•
1
LucasInsight/Meta-Llama-3-8B-Instruct
Shyamnath/Llama-3.2-3b-Uncensored-GGUF
Text Generation
•
Updated
•
17
•
3
ghost-x/ghost-8b-beta-1608-gguf
Text Generation
•
Updated
•
111
•
6
cahaj/Phi-3.5-mini-instruct-text2sql-GGUF
Updated
•
49
Agnuxo/Tinytron-Qwen-0.5B-Instruct_CODE_Python_Spanish_English_16bit
Updated
Agnuxo/Tinytron-Qwen-0.5B-TinyLlama-Instruct_CODE_Python-extra_small_quantization_GGUF_3bit
Updated
Agnuxo/Tinytron-Qwen-0.5B-Instruct_CODE_Python-Spanish_English_GGUF_4bit
Updated
Agnuxo/Tinytron-Qwen-0.5B-TinyLlama-Instruct_CODE_Python-Spanish_English_GGUF_q5_k
Updated
Agnuxo/Tinytron-Qwen-0.5B-TinyLlama-Instruct_CODE_Python-Spanish_English_GGUF_q6_k
Updated
Agnuxo/Tinytron-Qwen-0.5B-Instruct_CODE_Python-GGUF_Spanish_English_8bit
Updated
Agnuxo/Tinytron-Qwen-0.5B-Instruct_CODE_Python_English_GGUF_16bit
Agnuxo/Tinytron-Qwen-0.5B-TinyLlama-Instruct_CODE_Python-Spanish_English_GGUF_32bit
Updated