
A production-grade, Rust-powered AISIX gateway designed for performance, governance, and observability.
Blazing Fast, Built with Rust
Native Rust data plane delivers sub-millisecond proxy overhead with minimal memory footprint. Handle millions of requests per second without breaking a sweat.
One API, All Your LLMs
Manage all your LLM providers through a single, OpenAI-compatible API. Centralized configuration, authentication, and policy enforcement across your entire AI stack.
Multi-LLM Load Balancing
Dynamically distribute traffic across multiple LLM providers based on latency, cost, and availability. Weighted round-robin, least-connections, and custom strategies.
Token & Request Rate Limiting
Fine-grained rate limiting by tokens, requests, or custom dimensions. Per-consumer, per-route, and cluster-wide policies to control costs and prevent abuse.
Enterprise-Grade Security
Protect your AI pipeline with prompt injection detection, content moderation, PII redaction, and comprehensive audit logging for regulatory compliance.
Full-Stack Observability
Track every token, monitor latency distributions, and analyze traffic patterns in real-time. Native integration with Prometheus, Grafana, and ClickHouse.
Control Plane and Data Plane separation ensures high scalability, zero-downtime upgrades, and enterprise-level reliability.
Horizontally scalable with zero state — add or remove nodes instantly without data migration.
Centralized configuration management with real-time propagation and hot-reload capabilities.
Control Plane handles management; Data Plane handles traffic. Independent scaling and upgrades.
OpenAI-compatible API. Production-ready from day one.
Enterprise-grade security built into every request

Connect to any major LLM provider through a unified, OpenAI-compatible interface. No vendor lock-in, ever.
Start free with our open-source core. Scale to enterprise when you're ready.
Everything you need to get started with AISIX
Cloud or Self-Hosted. For teams managing AI traffic at scale.
Common questions from platform engineers evaluating AI gateways.
An AI gateway is a reverse proxy that sits between your applications and LLM providers (OpenAI, Anthropic, Google Gemini, DeepSeek, etc.). It centralizes authentication, load balancing, rate limiting, observability, and security policies for all AI and LLM traffic — similar to how a traditional API gateway manages REST API traffic.
Explore documentation, comparisons, and guides to evaluate AISIX for your AI infrastructure.
Quickstart guides, API references, deployment tutorials, and plugin documentation for the AISIX AI gateway.
Explore the AISIX source code, file issues, and contribute to the open-source AI gateway project.
A comprehensive guide to AI gateways: architecture patterns, LLM traffic governance, and production deployment strategies.
Understand the differences between AI gateways, MCP gateways, and traditional API gateways — and when to use each.
Deep dive into AI gateway architecture, use cases, and how it fits into your LLM infrastructure stack.
Compare AISIX against Kong, NGINX, Traefik, Tyk, and other API gateways across features, performance, and pricing.