LlamaLite-3B-TQ2_0 (GGUF Format)
This is a quantized version of meta-llama/Llama-3.2-3B-Instruct
, using TQ2_0 quantization for optimized performance and reduced size. The model is stored in GGUF format for compatibility with llama.cpp
and other lightweight inference engines.
Model Details
- Base Model: Llama-3.2-3B-Instruct
- Quantization Type:
TQ2_0
- Model Size: ~1.52GB
- Format: GGUF
- Intended Use: Text Generation, Chatbots, AI Assistants
- License: MIT
Download & Usage
1️⃣ Install Dependencies
pip install huggingface_hub
- Downloads last month
- 27
Hardware compatibility
Log In
to view the estimation
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for sagar27kumar/LlamaLite-3B-TQ2_0
Base model
meta-llama/Llama-3.2-3B-Instruct