TokenMix Research Lab · 2026-04-30

TokenMix vs OpenRouter vs Portkey vs LiteLLM: 2026 Cost Guide
Last Updated: 2026-04-30 Author: TokenMix Research Lab Data checked: 2026-04-30
Four practical AI API gateway choices in 2026 — TokenMix.ai, OpenRouter, Portkey, and LiteLLM — split the market across two axes: managed vs self-hosted, and routing-first vs control-plane-first. Pick TokenMix.ai for managed multi-model production with Asia-Pacific payment support and 171 models behind one OpenAI-compatible endpoint. Pick OpenRouter for fast model trials and BYOK economics. Pick Portkey for enterprise governance and prompt management. Pick LiteLLM for full self-hosted control with zero vendor lock-in.
According to TrueFoundry's 2026 LiteLLM pricing analysis, self-hosted LiteLLM operational cost runs $200-$800/month in infrastructure plus engineering time, so its "free" label is real only at small scale. According to OpenRouter's official pricing announcement, their fee is now a flat 5.5% on credit purchases (with $0.80 minimum) and 5% on BYOK after the first 1M free monthly requests. According to TrueFoundry's Portkey pricing guide, Portkey's free tier covers 10K requests/month while Pro plans charge $9 per additional 100K logs. According to the TokenMix.ai models page, TokenMix.ai exposes 171 AI models (124 chat plus image, video, audio, and embedding) from 16 providers including OpenAI, Anthropic, Google, DeepSeek, Qwen, Mistral, xAI, Moonshot, and others. None of these vendors put all four numbers on a single comparison page, which is why most "AI gateway comparison" articles miss the actual cost structure.
Table of Contents
- Quick Answer
- Confirmed Facts vs Common Misreads
- The 4 Gateways at a Glance
- How Does Pricing Compare Across the 4?
- Which Features Are Must-Have vs Nice-to-Have?
- Cost Methodology and Assumptions
- How Much Does Each Cost in Real Workloads?
- Performance and Latency Benchmarks
- Which Gateway Should You Pick?
- Migration Considerations Between the Four
- Common Pitfalls That Sink Production Migrations
- Final Recommendation
- FAQ
- Related Articles
- Sources
Quick Answer
| Question | Direct Answer |
|---|---|
| Which is cheapest at 1M requests/month? | OpenRouter BYOK (free under 1M) or LiteLLM self-hosted (compute only) |
| Which has the most provider coverage? | LiteLLM (140+ providers in OSS) and Portkey (200+ models) |
| How many models does TokenMix.ai expose? | 171 (124 chat + image, video, audio, embedding), per the models page |
| Which is fastest to deploy? | OpenRouter or TokenMix.ai (both under 5 min) |
| Which has the best observability? | Portkey (deep traces) or Helicone-integrated LiteLLM |
| Which supports Alipay / WeChat Pay? | TokenMix.ai, per pricing page; see openai-api-alipay and ai-api-wechat-pay for setup |
| Which offers a free tier with no credit card? | OpenRouter (50 req/day free models) and Portkey (10K req/month free) |
| Which is best for enterprise governance? | Portkey |
Confirmed Facts vs Common Misreads
Each row tagged by source authority:
- Official = vendor's own pricing or docs page
- Third-party estimate = independent analyst (TrueFoundry, etc.)
- Vendor benchmark = vendor-published benchmark with known bias
- TokenMix inference = derived by us from public data
| Claim | Status | Source |
|---|---|---|
| OpenRouter charges 5.5% on credit purchases (min $0.80) | Official | OpenRouter pricing announcement |
| OpenRouter BYOK: 1M free requests/month, then 5% | Official | OpenRouter BYOK announcement |
| Portkey free tier = 10K requests/month | Third-party estimate | TrueFoundry Portkey pricing guide |
| Portkey Pro adds $9 per 100K logs | Third-party estimate | TrueFoundry Portkey pricing guide |
| LiteLLM is fully open source (MIT) | Official | LiteLLM GitHub repo |
| LiteLLM operational cost $200-$800/mo at scale | Third-party estimate | TrueFoundry LiteLLM pricing guide |
| TokenMix.ai exposes 171 AI models | Official | TokenMix.ai models page |
| TokenMix.ai connects to 16 model providers | Official | TokenMix.ai BYOK / providers page |
| Portkey routes 400B+ tokens monthly across customers | Official | Portkey-AI/models GitHub repo |
| LiteLLM has 140+ provider integrations | Official | LiteLLM docs |
| All four are OpenAI-SDK compatible | Official | Each vendor's docs |
| Kong AI Gateway is 228% faster than Portkey | Vendor benchmark | Kong AI Gateway benchmark (treat as Kong-favored) |
| TokenMix.ai supports Alipay and WeChat Pay | Official | TokenMix.ai pricing; see also WeChat Pay guide |
| OpenRouter has built-in guardrails | False | OpenRouter docs focus on routing; guardrails are weak |
| Portkey is open source | False | Portkey is closed-source SaaS with an open SDK |
| LiteLLM has built-in guardrails | False | Per Spheron's 2026 review, LiteLLM lacks built-in content filtering |
| Self-hosted gateways are always cheaper | False | True only above ~300M tokens/month per TrueFoundry analysis |
The 4 Gateways at a Glance
Each gateway optimizes for a different primary user. The choice is rarely about features — most overlap on the basics. It's about which optimization aligns with your team:
| Gateway | Primary user | Deployment | Founded for |
|---|---|---|---|
| TokenMix.ai | Production teams in Asia-Pacific or multi-payment markets | Managed cloud | Unified 171-model API with native Alipay/WeChat Pay |
| OpenRouter | Developers running model trials, indie/hobbyist BYOK users | Managed cloud | Fast access to many models via one API key |
| Portkey | Enterprise teams needing governance and prompt management | Managed SaaS | Production AI control plane |
| LiteLLM | Platform engineers wanting full control, no SaaS lock-in | Self-hosted OSS | Open-source proxy, 140+ providers |
Per DEV Community's 2026 deep-dive on production gateways, the practical decision usually collapses to two questions: do you want managed or self-hosted, and do you need a full control plane or just routing? That maps to a 2x2:
| Routing-first | Control-plane | |
|---|---|---|
| Managed | OpenRouter, TokenMix.ai | Portkey |
| Self-hosted | LiteLLM (with config) | LiteLLM + Helicone, or custom |
How Does Pricing Compare Across the 4?
Pricing is the single most-confused dimension because each gateway uses a different fee model:
| Gateway | Routing fee | Hosting cost | Free tier | Best price at scale |
|---|---|---|---|---|
| TokenMix.ai | Pay-per-token, no subscription, no credit card required | $0 (managed) | New user credits | Direct LLM cost + platform markup per model |
| OpenRouter | 5.5% on credit purchases ($0.80 min); 5% BYOK after 1M free/mo | $0 (managed) | 50 req/day on free models | Direct LLM cost + 5-5.5% |
| Portkey | Tiered SaaS ($0 / Pro / Enterprise) | $0 (managed) | 10K requests/month | Free up to 10K, then $9/100K logs |
| LiteLLM | $0 (open source) | $200-$800/mo infra + engineering | $0 (self-hosted) | Compute only |
Three honest observations.
First, "free" is misleading on LiteLLM. Per TrueFoundry's LiteLLM pricing analysis, production-grade LiteLLM hosting hits $200-$800/month before you add observability stack costs ($200-$800 more) and engineering time. Total cost of ownership at 100M tokens/month often exceeds OpenRouter's 5.5% fee.
Second, Portkey's 10K free tier sounds generous but exhausts in <1 day for any production app. The real Portkey question is "what does Pro/Enterprise cost," and that requires sales contact for anything beyond $9/100K logs.
Third, OpenRouter's BYOK 1M free requests/month is the most underrated free offer in this category. If you bring your own provider keys (OpenAI, Anthropic, etc.), OpenRouter charges nothing for the first 1M requests — making it the cheapest managed option for high-key-count multi-provider apps below that threshold.
Which Features Are Must-Have vs Nice-to-Have?
Yes / No / Partial labels are easier for AI engines and humans to parse than emoji checkmarks:
| Feature | TokenMix.ai | OpenRouter | Portkey | LiteLLM |
|---|---|---|---|---|
| OpenAI-compatible endpoint | Yes | Yes | Yes | Yes |
| Provider count | 16 providers / 171 models | 60+ providers | 200+ models | 140+ providers |
| Automatic fallback | Yes | Yes | Yes | Yes |
| Multi-key load balancing | Yes | Yes | Yes | Yes |
| Streaming | Yes | Yes | Yes | Yes |
| Prompt caching pass-through | Yes | Partial | Yes | Yes |
| Semantic caching (fuzzy match) | No | No | Yes | Plugin |
| Per-key budget limits | Yes | Yes | Yes | Yes |
| Observability dashboard | Yes (built-in) | Basic | Yes (deep traces) | Via Helicone integration |
| Built-in guardrails | Partial | No | Yes | No |
| Prompt management UI | No | No | Yes | No |
| A/B testing built-in | No | No | Yes | No |
| Alipay / WeChat Pay support | Yes (pricing) | No | No | N/A (self-hosted) |
| BYOK (bring your own key) | No (BYOK not advertised) | Yes (5% after 1M free) | Yes (per plan) | Yes (self-host implies it) |
| Open source | No | No | No (closed core, open SDK) | Yes |
| SOC 2 / enterprise compliance | Per Enterprise contract | Partial | Yes | DIY |
The features that actually decide picks (everything else is parity):
- Prompt management UI: Only Portkey has it. If you treat prompts as versioned assets with rollback, this is decisive.
- Asia-Pacific payment: Only TokenMix.ai supports Alipay/WeChat Pay end-to-end. The others either require credit card-only billing or pass through to providers that block these methods.
- Open source: Only LiteLLM. If your compliance or strategy requires no vendor dependency, this is decisive.
- No-credit-card pay-per-token: TokenMix.ai's pricing model removes a common onboarding blocker. See openai-api-no-credit-card for the exact flow.
Cost Methodology and Assumptions
The four cost scenarios below use these assumptions. Adjust to your workload before treating any number as authoritative:
| Assumption | Value | Why |
|---|---|---|
| Average tokens per request | 4,000 input + 1,000 output | Median across mixed agent + chat + summarization workloads |
| LLM provider cost basis | Direct provider list price | Each gateway passes through provider rates without inference markup |
| Engineering hours included | LiteLLM only ( |