LLM / VLM Quantization
Collection
3 items
•
Updated
prepared with mergekit
quantized 4 bits with bitsandbytes
base_model: meta-llama/Llama-2-7b-chat-hf
gate_mode: cheap_embed
experts:
- source_model: meta-llama/Llama-2-7b-chat-hf
positive_prompts: ["You are an helpful assistant."]
- source_model: TheTravellingEngineer/llama2-7b-hf-guanaco
positive_prompts: ["You are an helpful general-pupose assistant"]