🚀 Feature Description and Motivation
I’m currently trying to integrate a mock Aibrix-deployed model (e.g., llama2-7b) with Envoy AI Gateway using the Kubernetes Gateway API and the InferencePool / AIGatewayRoute pattern (similar to the official envoyproxy/ai-gateway ai-gateway).
Use Case
We want to expose multiple Aibrix-managed LLMs (e.g., llama2-7b, qwen-14b) through a single, secure, OpenAI-compatible API endpoint powered by Envoy AI Gateway and gateway-api-inference-extension
Proposed Solution
🚀 Feature Description and Motivation
I’m currently trying to integrate a mock Aibrix-deployed model (e.g., llama2-7b) with Envoy AI Gateway using the Kubernetes Gateway API and the InferencePool / AIGatewayRoute pattern (similar to the official envoyproxy/ai-gateway ai-gateway).
Use Case
We want to expose multiple Aibrix-managed LLMs (e.g., llama2-7b, qwen-14b) through a single, secure, OpenAI-compatible API endpoint powered by
Envoy AI Gatewayandgateway-api-inference-extensionProposed Solution