Edge AI Inference Gateway
A globally distributed inference gateway running on Cloudflare Workers, routing requests to the nearest Workers AI node. Supports streaming, model fallbacks, and per-route rate limiting with sub-20ms overhead at the 95th percentile.