Skip to content

Conversation

@AviadHayumi
Copy link

Previously, MODELS env var could only override models discovered from OPENAI_BASE_URL. This change allows MODELS entries with explicit endpoints to be added as new models when they don't match any discovered model.

This enables configuring multiple LLM backends (e.g., NIM + FakeLLM) from different endpoints without requiring a proxy or load balancer.

Previously, MODELS env var could only override models discovered from
OPENAI_BASE_URL. This change allows MODELS entries with explicit
endpoints to be added as new models when they don't match any
discovered model.

This enables configuring multiple LLM backends (e.g., NIM + FakeLLM)
from different endpoints without requiring a proxy or load balancer.

Example MODELS config for multiple backends:
[
  {"id":"model-from-api1","name":"Model 1","endpoints":[{"type":"openai","baseURL":"http://api1/v1"}]},
  {"id":"model-from-api2","name":"Model 2","endpoints":[{"type":"openai","baseURL":"http://api2/v1"}]}
]
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant