-
-
Notifications
You must be signed in to change notification settings - Fork 3.8k
Changing context_size in YAML Configuration Editor requires restarting #8647
Description
LocalAI version:
v.3.12.1
Environment, CPU architecture, OS, and Version:
docker
Describe the bug
When changing the model .yaml file via the frontend to change the context_size: 16384, after clicking the "Update" button, the models appear to reload but the change in the yaml is not used. If I simply restart the LocalAI docker container, the change is used. I believe this used to work as expected in a prior version of LocalAI.
To Reproduce
Make a change in the model's context_size in yaml via the frontend's "YAML configuration editor" and click save. Note the change is not implemented.
Expected behavior
changed configuration.
Logs
Config:
backend: llama-cpp
context_size: 16384
f16: true
mmap: true
mmproj: mmproj/mmproj-Qwen3-VL-2B-Instruct-F16.gguf
name: qwen3-vl-2b-instruct
parameters:
model: Qwen3-VL-2B-Instruct-Q4_K_M.gguf
...
Log:
Feb 24 14:02:53 DEBUG GRPC stderr id="qwen3-vl-2b-instruct-127.0.0.1:34731" line="srv send_error: task id = 321, error: request (8899 tokens) exceeds the available context size (8192 tokens), try increasing it" caller={caller.file="/build/pkg/model/process.go" caller.L=153 }
Additional context