runtime error
Exit code: 3. Reason: > n_ctx_train (4096) -- possible training context overflow set_abort_callback: call llama_context: CUDA_Host output buffer size = 0.12 MiB llama_kv_cache: CUDA0 KV buffer size = 3300.00 MiB llama_kv_cache: size = 3300.00 MiB ( 4224 cells, 40 layers, 1/1 seqs), K (f16): 1650.00 MiB, V (f16): 1650.00 MiB llama_context: enumerating backends llama_context: backend_ptrs.size() = 2 llama_context: max_nodes = 2904 llama_context: reserving full memory module llama_context: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 llama_context: CUDA0 compute buffer size = 123.39 MiB llama_context: CUDA_Host compute buffer size = 18.26 MiB llama_context: graph nodes = 1247 llama_context: graph splits = 2 attach_threadpool: call clip_model_loader: model name: openai/clip-vit-large-patch14-336 clip_model_loader: description: image encoder for LLaVA clip_model_loader: GGUF version: 3 clip_model_loader: alignment: 32 clip_model_loader: n_tensors: 377 clip_model_loader: n_kv: 19 clip_model_loader: has vision encoder clip_ctx: CLIP using CUDA0 backend clip_init: failed to load model '/opt/koboldcpp/mmproj.gguf': load_hparams: unknown projector type: gpttype_load_model: error: failed to load mmproj model! Load Text Model OK: False Could not load text model: /opt/koboldcpp/model.gguf The reported GGUF Arch is: llama Arch Category: 0 --- Identified as GGUF model. Attempting to Load... --- Using automatic RoPE scaling for GGUF. If the model has custom RoPE settings, they'll be used directly instead! System Info: AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | AMX_INT8 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | RISCV_VECT = 0 | WASM_SIMD = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | CUDA MMQ: False Automatic RoPE Scaling: Using (scale:1.000, base:10000.0). Threadpool set to 2 threads and 2 blasthreads... Attempting to apply Multimodal Projector: /opt/koboldcpp/mmproj.gguf
Container logs:
Fetching error logs...