runtime error

Exit code: 1. Reason: load_backend: loaded CPU backend from /app/libggml-cpu-icelake.so build: 6445 (00681dfc) with cc (Ubuntu 11.4.0-1ubuntu1~22.04.2) 11.4.0 for x86_64-linux-gnu system info: n_threads = 8, n_threads_batch = 8, total_threads = 16 system_info: n_threads = 8 (n_threads_batch = 8) / 16 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | BMI2 = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | LLAMAFILE = 1 | OPENMP = 1 | REPACK = 1 | main: binding port with default address family main: HTTP server is listening, hostname: 0.0.0.0, port: 7860, http threads: 15 main: loading model srv load_model: loading model 'coconut-qwen2.5-7b.Q4_K_S.gguf' gguf_init_from_file: failed to open GGUF file 'coconut-qwen2.5-7b.Q4_K_S.gguf' llama_model_load: error loading model: llama_model_loader: failed to load model from coconut-qwen2.5-7b.Q4_K_S.gguf llama_model_load_from_file_impl: failed to load model common_init_from_params: failed to load model 'coconut-qwen2.5-7b.Q4_K_S.gguf', try reducing --n-gpu-layers if you're running out of VRAM srv load_model: failed to load model, 'coconut-qwen2.5-7b.Q4_K_S.gguf' srv operator(): operator(): cleaning up before exit... main: exiting due to model loading error

Container logs:

Fetching error logs...