runtime error

Exit code: 1. Reason: merged_model-q4.gguf: 0%| | 0.00/4.22G [00:00<?, ?B/s] merged_model-q4.gguf: 0%| | 10.5M/4.22G [00:01<09:30, 7.38MB/s] merged_model-q4.gguf: 3%|▎ | 115M/4.22G [00:02<01:13, 56.0MB/s]  merged_model-q4.gguf: 9%|▉ | 398M/4.22G [00:03<00:26, 146MB/s]  merged_model-q4.gguf: 23%|██▎ | 965M/4.22G [00:04<00:10, 299MB/s] merged_model-q4.gguf: 33%|███▎ | 1.41G/4.22G [00:05<00:08, 346MB/s] merged_model-q4.gguf: 42%|████▏ | 1.77G/4.22G [00:06<00:07, 328MB/s] merged_model-q4.gguf: 55%|█████▍ | 2.31G/4.22G [00:07<00:04, 390MB/s] merged_model-q4.gguf: 64%|██████▍ | 2.72G/4.22G [00:09<00:04, 369MB/s] merged_model-q4.gguf: 74%|███████▎ | 3.11G/4.22G [00:10<00:02, 377MB/s] merged_model-q4.gguf: 86%|████████▌ | 3.62G/4.22G [00:11<00:01, 413MB/s] merged_model-q4.gguf: 100%|█████████▉| 4.22G/4.22G [00:12<00:00, 345MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 18, in <module> model, tokenizer = load_model() File "/home/user/app/app.py", line 12, in load_model model = AutoModelForCausalLM.from_pretrained(model_path, model_type="llama") File "/usr/local/lib/python3.10/site-packages/ctransformers/hub.py", line 175, in from_pretrained llm = LLM( File "/usr/local/lib/python3.10/site-packages/ctransformers/llm.py", line 253, in __init__ raise RuntimeError( RuntimeError: Failed to create LLM 'llama' from '/home/user/.cache/huggingface/hub/models--Tommi09--MedicalChatBot-7b-test/snapshots/9850350e1353823602e5127325692c8104240c23/LoRA-Huatuo-7b-GGUF-Q4/merged_model-q4.gguf'.

Container logs:

Fetching error logs...