Pipelex Gateway & Model Access
Access AI models through the Pipelex Gateway or bring your own API keys.
Pipelex Gateway
A fully managed infrastructure providing unified access to AI models through a single API key. The Gateway eliminates the need to manage multiple provider configurations.
- Single API key for all supported models
- Remote model catalog — always access the latest models without updating Pipelex
- Enterprise-grade architecture — built for reliability and scale
- Extensive provider support — OpenAI, Google, Anthropic, Mistral, xAI, and more
Browse all supported models in the Gateway Models reference.
Get your Gateway API key at app.pipelex.com or join the waitlist.
Bring Your Own Keys
Direct integration with major providers using your own API keys:
- OpenAI — GPT-4o, GPT-4.1, o1, o3, o4-mini, etc.
- Anthropic — Claude Sonnet 4, Claude Haiku, etc.
- Google — Gemini 2.5 Pro, Gemini 2.5 Flash, etc.
- Mistral — Mistral Large, Mistral Medium, etc.
- Azure OpenAI — Azure-hosted OpenAI models
- Amazon Bedrock — AWS-hosted models
- Vertex AI — Google Cloud-hosted models
- xAI — Grok models
- Portkey — AI gateway for routing and observability
- OpenRouter — Multi-provider aggregator
- BlackboxAI — Blackbox-hosted models
- fal — Image generation models
Open-Source Models
Run open-source models through dedicated providers:
- Ollama — Run open-source LLMs locally (Llama, Mistral, Qwen, etc.)
- Hugging Face Inference — including qwen-image for text-to-image
- Scaleway — Deepseek R1, Llama 3.3, Qwen3, GPT-OSS
- Groq — Llama-4, Kimi-K2-Instruct
Routing Profiles
Routing profiles control which backend handles each model. Define pattern-based routes, defaults, and fallback orders in routing_profiles.toml. Switch a pipeline from one provider to another — or from Pipelex Gateway to your own keys — without changing the method definition.
See Configure AI Providers for setup details.