Skip to main content

2 posts tagged with "Features"

Articles about Envoy AI Gateway Features

View All Tags

Enhancing AI Gateway Observability - OpenTelemetry Tracing Arrives in Envoy AI Gateway

· 6 min read
Erica Hughberg
Envoy AI Gateway Maintainer - Tetrate
Adrian Cole
Principal Engineer - Tetrate

Hero feature image of title.

Aggregated metrics like latency, error rates, and throughput on their own won't reveal the source of why a system's output was wrong, slow, or expensive.

The v0.3 release of Envoy AI Gateway brings comprehensive OpenTelemetry tracing support with OpenInference semantic conventions, extending the existing metrics foundation to provide complete visibility into LLM application behavior.

This enables you to improve the quality and safety of your AI-integrated applications by allowing you to understand the full context of a request journey, as your LLM traces will inform application improvements and guardrail needs.

Envoy AI Gateway Introduces Endpoint Picker Support

· 7 min read
Erica Hughberg
Envoy AI Gateway Maintainer - Tetrate
Xunzhuo (Bit) Liu
Envoy AI Gateway Maintainer - Tencent

Reference Architecture for Envoy AI Gateway

Introduction

Envoy AI Gateway now supports Endpoint Picker Provider (EPP) integration as per the Gateway API Inference Extension.

This feature enables you to leverage intelligent, dynamic routing for AI inference workloads through intelligent endpoint selection based on real-time metrics, including KV-cache usage, queued requests, and LoRA adapter information.

When running AI inference at scale, this means your system can automatically select the optimal inference endpoint for each request, thereby optimizing resource utilization.

An overview of Endpoint Picker together with Envoy AI Gateway