Skip to content

Changing context_size in YAML Configuration Editor requires restarting #8647

@loopy321

Description

@loopy321

LocalAI version:

v.3.12.1

Environment, CPU architecture, OS, and Version:

docker

Describe the bug

When changing the model .yaml file via the frontend to change the context_size: 16384, after clicking the "Update" button, the models appear to reload but the change in the yaml is not used. If I simply restart the LocalAI docker container, the change is used. I believe this used to work as expected in a prior version of LocalAI.

To Reproduce

Make a change in the model's context_size in yaml via the frontend's "YAML configuration editor" and click save. Note the change is not implemented.
Expected behavior

changed configuration.

Logs

Config:

backend: llama-cpp
context_size: 16384
f16: true
mmap: true
mmproj: mmproj/mmproj-Qwen3-VL-2B-Instruct-F16.gguf
name: qwen3-vl-2b-instruct
parameters:
    model: Qwen3-VL-2B-Instruct-Q4_K_M.gguf
...

Log:
Feb 24 14:02:53 DEBUG GRPC stderr id="qwen3-vl-2b-instruct-127.0.0.1:34731" line="srv send_error: task id = 321, error: request (8899 tokens) exceeds the available context size (8192 tokens), try increasing it" caller={caller.file="/build/pkg/model/process.go" caller.L=153 }

Additional context

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions