Curated Eval Suites
10+ pre-built evaluation suites covering safety, quality, RAG faithfulness, and compliance. Start evaluating in minutes with battle-tested benchmarks.
Evaluate continuously, route intelligently, comply automatically. The first AI gateway where eval results drive every decision.
Most evaluation tools are disconnected from production. You run evals in notebooks, see results in dashboards, but nothing changes automatically. GateFlow closes the loop.
GateFlow enables quality-driven AI infrastructure:
Evals run automatically on production traffic:
from openai import OpenAI
client = OpenAI(
base_url="https://api.gateflow.ai/v1",
api_key="gf-..." # Your GateFlow API key
)
# Standard inference - evals sample automatically
response = client.chat.completions.create(
model="auto", # Routing informed by eval scores
messages=[{"role": "user", "content": "Hello!"}]
)
# Or run explicit eval suites
from gateflow import EvalClient
eval_client = EvalClient(api_key="gf-...")
results = eval_client.run_suite(
suite="safety-core",
model="gpt-4o"
)
print(f"Safety score: {results.aggregate_score}%")That's it. Your requests flow through GateFlow with automatic evaluation, quality-driven routing, and compliance reporting.
Latest Release - v3.0.0
Eval Platform Launch - Gateway-native evaluation with 10+ curated suites, tiered evaluators for 97% cost reduction, closed-loop routing, and EU AI Act compliance reporting. Read the changelog →
Eval Platform
Production Integration
Compliance
Gateway Features