Skip to content

Best Open Source LLM Gateways of 2026

Updated · 3 picks · live pricing · affiliate disclosure

Apache 2.0 production gateway since March 2026 with caching, fallbacks, prompt management, and guardrails.

BEST OVERALL5.8/10$240/yr more

Portkey

Apache 2.0 production gateway since March 2026 with caching, fallbacks, prompt management, and guardrails.

OSS Apache 2.0 free forever; Developer Free 10K logs/mo

How it stacks up

  • OSS Apache 2.0 free self-host

    vs LiteLLM focused proxy

  • Production $49/mo upgrade

    vs Langfuse observability

  • Caching + guardrails

    OSS production gateway

#2
LiteLLM5.5/10

From $50/mo

View
#3
Langfuse4.3/10

From $29/mo

View

All picks at a glance

#PickBest forStartingScore
1PortkeyBest OSS production gateway, Apache 2.0 with caching and guardrails$49.00/mo5.8/10
2LiteLLMBest OSS LLM proxy, Apache 2 / MIT with 100+ providers$50.00/mo5.5/10
3LangfuseBest OSS observability, MIT-licensed self-host with traces and evals$29.00/mo4.3/10

Quick pick by use case

If you only have thirty seconds, find your situation below and skip to that pick.

Compare all 3 picks

Top spec
#1Portkey5.8/10$49.00/mo$588.00/yr$240/yr moreOSS Apache 2.0 free self-host
#2LiteLLM5.5/10$50.00/mo$600.00/yr$252/yr moreOSS Apache 2 / MIT free
#3Langfuse4.3/10$199.00/mo$2,388.00/yr$2,040/yr moreOSS MIT free self-host
#1

Portkey

5.8/10$240/yr more

Best OSS production gateway, Apache 2.0 with caching and guardrails

Apache 2.0 production gateway since March 2026 with caching, fallbacks, prompt management, and guardrails.

PlanMonthlyAnnualWhat you get
Open SourceFreeApache 2.0 self-hosted gateway with Universal API, retries, routing, guardrails, automatic fallbacks, basic dashboard, and load balancing; free forever (open-sourced March 2026)
Developer FreeFreeFree hosted tier with 10K recorded logs a month, Universal API, key management, 3 prompt templates, and basic observability
Production$49.00/mo$588.00/yr$49 a month for 100K logs with unlimited prompt templates, alerts, LLM guardrails, semantic caching, RBAC, and overages at $9 per additional 100K (up to 3M)
Enterprise$2,000.00/mo$24,000.00/yrCustom contract with 10M-plus logs, custom retention, advanced guardrails, SSO, VPC hosting, SOC 2 Type II, GDPR, and HIPAA

Portkey is the OSS production-gateway pick that fully open-sourced under Apache 2.0 on March 24, 2026. Founded 2023 in San Francisco. The wedge: Apache 2.0 self-host ships caching, fallbacks, prompt management, guardrails, RBAC, and observability across 40-plus providers; the same feature set that production teams pay for on Cloud is available free for self-host. Cloud handles around one trillion tokens a day across roughly 24,000 organizations.

Open Source Apache 2.0 is free forever for self-host with Universal API, retries, routing, guardrails, fallbacks, and load balancing. Developer Free hosted covers 10,000 logs a month. Production is the upgrade tier at forty-nine dollars monthly with 100,000 logs, unlimited prompt templates, semantic caching, LLM guardrails, and RBAC. Enterprise covers VPC hosting, SSO, SOC 2 Type II, GDPR, and HIPAA.

The trade-off versus LiteLLM is scope; Portkey is a full production gateway where LiteLLM is a focused routing proxy. The trade-off versus Langfuse is layer; Portkey bundles routing with observability where Langfuse is observability-only. For OSS production teams who want caching and guardrails on one self-hosted gateway, Portkey is the right call.

Pros

  • Apache 2.0 OSS gateway since March 2026; production feature set free for self-host
  • Caching, fallbacks, prompt management, guardrails, and observability on one dashboard
  • Cloud handles around 1 trillion tokens daily across roughly 24,000 organizations
  • Production upgrade at forty-nine dollars monthly for managed hosting
  • Enterprise covers VPC hosting with SOC 2 Type II, GDPR, and HIPAA

Cons

  • Self-host requires you to run and monitor the gateway versus managed Cloud
  • Younger OSS community than LiteLLM at this stage
OSS Apache 2.0 free self-hostProduction $49/mo upgradeCaching + guardrailsOSS Apache 2.0 free forever; Developer Free 10K logs/mo

Best for: OSS production teams who want caching, guardrails, and observability on one self-hosted Apache 2.0 gateway with full feature parity to managed Cloud.

Routing
9
Latency
9
DX
8
Value
8
Support
8
#2

LiteLLM

5.5/10$252/yr more

Best OSS LLM proxy, Apache 2 / MIT with 100+ providers

Apache 2 / MIT Python proxy translating 100+ provider APIs to OpenAI Chat Completions schema; YC W23.

PlanMonthlyAnnualWhat you get
Open SourceFreeApache 2 / MIT-licensed self-hosted Python proxy that translates 100+ provider APIs to the OpenAI Chat Completions schema; free forever
Cloud FreeFreeFree hosted tier with limited request volume and standard provider integrations
Cloud Pro$50.00/mo$600.00/yr$50 per user a month with cost tracking, budgets, team management, and virtual API keys
Enterprise$2,000.00/mo$24,000.00/yrCustom contract with self-hosted enterprise, SSO, SOC 2, and HIPAA available

LiteLLM is the Apache 2 / MIT-licensed OSS proxy pick and the largest community among self-hosted LLM gateway proxies. Founded 2023 by BerriAI Inc (Y Combinator W23) in San Francisco. The wedge: Apache 2 / MIT permissive license, Python proxy that translates 100-plus provider APIs to the OpenAI Chat Completions schema, drop-in replacement at the application level so OpenAI SDK calls work unchanged.

Open Source is Apache 2 / MIT-licensed and free forever for self-hosted deployment with 100-plus provider integrations. Cloud Free is a hosted tier with limited request volume. Cloud Pro is the upgrade tier at fifty dollars per user monthly with cost tracking, budgets, team management, and virtual API keys. Enterprise covers self-hosted enterprise with SSO, SOC 2, and HIPAA.

The trade-off versus Portkey is scope; LiteLLM is a focused routing proxy where Portkey is a full production gateway with observability and guardrails. The trade-off versus Langfuse is layer; LiteLLM is gateway routing where Langfuse is observability. For OSS Python-stack readers who want a routing proxy with broad provider coverage, LiteLLM is the right call.

Pros

  • Apache 2 / MIT permissive license; all commercial use allowed at any revenue level
  • 100+ provider integrations free forever for self-hosted deployment
  • Drop-in replacement: OpenAI SDK calls work unchanged behind the proxy
  • Y Combinator W23; the most-recognized OSS LLM proxy among Python developers
  • Cloud Pro at fifty dollars per user for cost tracking and virtual API keys

Cons

  • Self-hosted Python proxy requires Python operational maturity to run reliably
  • Python overhead can add hundreds of microseconds at high concurrency versus Go OSS
OSS Apache 2 / MIT freeCloud Pro $50/user/mo100+ provider integrationsOSS Apache 2 / MIT free forever; cancel Cloud anytime

Best for: OSS Python-stack readers who want a routing proxy with broad provider coverage and accept Python self-host operational tax for full code control.

Routing
10
Latency
8
DX
7
Value
10
Support
7
#3

Langfuse

4.3/10$2,040/yr more

Best OSS observability, MIT-licensed self-host with traces and evals

MIT-licensed LLM observability with traces, evals, and prompt management; self-host option; YC W23 in Berlin.

PlanMonthlyAnnualWhat you get
HobbyFreeFree 50K observations a month with all platform features (limits apply), 30-day data retention, and 2 users
Core$29.00/mo$348.00/yr$29 a month for 100K observations with unlimited users, 90-day retention, and in-app support
Pro$199.00/mo$2,388.00/yr$199 a month with unlimited history, 3-year data access, SOC 2, ISO 27001, HIPAA, and 20K req/min throughput
Enterprise$2,499.00/mo$29,988.00/yr$2,499 a month with audit logs, SCIM API, custom rate limits, dedicated support engineer, and uptime SLAs

Langfuse is the MIT-licensed OSS observability pick and the most-popular OSS LLM observability project. Founded 2023 (Y Combinator W23) in Berlin Germany. Around 6,000 GitHub stars by 2026. The wedge: MIT-licensed observability with traces, evals, prompt management, and analytics on a project that you can run on your own infrastructure or use as Cloud.

Hobby is free with 50,000 observations a month, all platform features (with limits), 30-day data retention, and 2 users. Core is the upgrade tier at twenty-nine dollars monthly with 100,000 observations and unlimited users plus 90-day retention. Pro at the next upgrade adds 3-year retention, SOC 2, ISO 27001, HIPAA, and 20K req/min throughput. Enterprise covers audit logs, SCIM API, dedicated support, and SLAs.

The trade-off versus Helicone is gateway routing; Langfuse is observability-only where Helicone bundles gateway routing with observability. The trade-off versus LiteLLM is layer; Langfuse is the observability layer where LiteLLM is the proxy layer. For OSS readers who want self-hostable observability with unlimited observation volume, Langfuse is the right call. Pair with LiteLLM as the proxy layer for a full OSS gateway plus observability stack.

Pros

  • MIT permissive license; all commercial use allowed at any revenue level
  • Self-host option for unlimited observation volume on your own infrastructure
  • Hobby free covers 50,000 observations a month with all features and 2 users
  • Around 6,000 GitHub stars by 2026; the most-popular OSS LLM observability
  • Berlin-based with default EU data residency on Cloud

Cons

  • Observability-only; gateway routing requires a separate proxy layer (LiteLLM, OpenRouter)
  • Smaller community than LiteLLM at this stage
OSS MIT free self-hostHobby free 50K obs/moCore $29/mo upgradeHobby free 50K obs/mo; MIT OSS self-host free forever

Best for: OSS readers who want MIT-licensed observability they can self-host with unlimited observation volume; pair with LiteLLM for full OSS stack.

Routing
10
Latency
8
DX
8
Value
9
Support
7

How we picked

Each pick gets a transparent composite score from price, features, free-tier availability, and editor fit. Pricing flows from our live database, so when a vendor changes prices the score updates here too.

We weight price at 40 percent, features at 30, free tier at 15, fit at 15. LiteLLM leads because Apache 2 / MIT license plus 100-plus provider integrations plus the largest OSS proxy community is the strongest free-forever defensibility. See the parent /best/llm-gateways guide for closed-source SaaS picks excluded from this lens.

We don't claim "30,000 hours of testing." Our methodology is the formula above plus the editor's published verdict for each pick. Verifiable, auditable, and updated when the underlying data changes.

Why trust Subrupt

We're a subscription tracker first, a buying guide second. Every claim on this page is something you can check.

By use case

Best OSS proxy with 100+ providers

LiteLLM

Read the full review →

Best OSS production gateway

Portkey

Read the full review →

Best OSS observability

Langfuse

Read the full review →

How to choose your Open Source LLM Gateway

Apache 2 / MIT vs Apache 2.0 vs source-available

All three OSS picks here ship permissive licenses. LiteLLM is Apache 2 / MIT-licensed (permissive at any revenue level). Portkey open-sourced their full gateway under Apache 2.0 on March 24, 2026; permissive at any revenue level. Langfuse is MIT-licensed (permissive at any revenue level). Out-of-catalog OSS projects worth knowing include Bifrost (Apache 2.0 Go-based with around 11 microseconds overhead at 5,000 requests per second), Kong AI Gateway (Apache 2.0 enterprise gateway), and Apache APISIX (Apache 2.0 with multi-LLM load balancing). All four catalog plus out-of-catalog OSS picks permit commercial use including SaaS resale at any revenue level; license posture is uniform across the OSS lens. The decision pivots on architectural fit rather than license restrictions.

Layer split: proxy vs full gateway vs observability

The three catalog OSS picks divide cleanly across layers. LiteLLM is a routing proxy; you put it in front of your model providers and your application code targets the OpenAI SDK only. Portkey is a full production gateway with caching, fallbacks, guardrails, RBAC, and observability bundled. Langfuse is pure observability; you point your proxy or application at it and get traces, evals, and analytics without gateway routing. Many production systems run LiteLLM as the proxy layer plus Langfuse as the observability layer, both self-hosted, with zero per-request fees. Portkey bundles both layers in one product; the choice is whether you want one tool or two specialized tools at the OSS layer.

Self-host operational tax across the three picks

Self-host is dollar-zero on licensing but readers pay infrastructure plus operational labor. LiteLLM self-host runs on a single VPS at around five to twenty dollars monthly with the Python runtime; the operational tax is Python dependency management, LLM API key rotation, and proxy runtime monitoring. Portkey self-host runs on similar VPS infrastructure with Node.js; the operational tax is gateway dashboard monitoring and managing the broader feature set. Langfuse self-host runs on Postgres plus the Langfuse Node.js application; the operational tax is database backups, migrations, and observability ingestion scaling. For teams without DevOps capacity, the Cloud upgrades trade operational labor for managed hosting at fifty to seventy-nine dollars monthly across the three picks.

Out-of-catalog OSS gateways worth knowing

OSS LLM gateway lists frequently include projects we do not yet catalog. Bifrost is Apache 2.0 Go-based with around 11 microseconds gateway overhead at 5,000 requests per second under sustained traffic; competes with LiteLLM at the proxy layer with stronger production performance. Kong AI Gateway is Apache 2.0 enterprise gateway built on Kong's API gateway infrastructure; competes with Portkey at the full-gateway layer for orgs already running Kong. Apache APISIX is Apache 2.0 with multi-LLM load balancing, retry and fallback mechanisms, token rate limiting, and content moderation. From our catalog, LiteLLM, Portkey, and Langfuse are the OSS LLM gateway picks. Readers who specifically need Bifrost (Go-performance), Kong (enterprise), or APISIX (load-balancing) should evaluate those projects directly.

When to consider closed-source SaaS picks (cross-link to parent)

OSS self-host is the right path for teams that value auditable source code and accept the operational tax. Some teams have requirements that closed-source SaaS picks cover better. Indie-dev brand reference with 300-plus models is uniquely OpenRouter (closed-source SaaS); none of the OSS picks here ship that scale. Edge-native delivery across 300-plus data centers is Cloudflare AI Gateway (closed-source SaaS). Tight Vercel AI SDK integration is Vercel AI Gateway (closed-source SaaS). Observability-first one-line proxy with HQL query language is Helicone (closed-source SaaS). At any of those decision points, see [our /best/llm-gateways guide](/best/llm-gateways) for the closed-source picks excluded under the OSS lens.

Frequently asked questions

Why is LiteLLM ranked first over Portkey?

LiteLLM wins on community defensibility because Apache 2 / MIT license plus 100-plus provider integrations plus around 15,000 GitHub stars is the strongest free-forever profile in OSS LLM proxies. Portkey wins on production feature breadth specifically. Teams that want a focused routing proxy pick LiteLLM. Teams that want a full production gateway with caching and guardrails pick Portkey. Both are auditable OSS at any revenue level.

Does Subrupt earn a commission from these OSS picks?

On a few. We disclose this on every /best page. Self-host OSS (LiteLLM Apache 2 / MIT, Portkey Apache 2.0, Langfuse MIT) has no affiliate path because there is no transaction. Cloud upgrades on LiteLLM Pro, Portkey Production, and Langfuse Core have paid plans where we earn commission only on conversion. The composite ranking weights price at 40 percent, features at 30, free tier at 15, fit at 15; none tuned by affiliate rate.

Should I run LiteLLM plus Langfuse together or use Portkey for both?

LiteLLM plus Langfuse splits the proxy and observability layers; you run two specialized tools that each excel at their layer. Portkey bundles both in one product. The decision pivots on team posture. Teams that want maximum flexibility and accept operational tax for two services pick the LiteLLM plus Langfuse stack. Teams that want one tool covering both layers pick Portkey. Both paths are dollar-zero on licensing.

How much does self-hosting actually cost in infrastructure?

A small LiteLLM deployment runs on a single five to fifteen dollar VPS with Python. A small Portkey OSS deployment runs on a similar VPS with Node.js. Langfuse self-host needs Postgres plus the Langfuse application; around fifteen to thirty dollars monthly. For staging and prod, double the cost. Above that, you absorb the operational tax of patching, backups, monitoring, and on-call. Cloud upgrades trade labor for managed hosting at fifty to seventy-nine dollars monthly.

Is Bifrost actually faster than LiteLLM?

Yes for high-concurrency production workloads. Bifrost is Go-based with around 11 microseconds gateway overhead at 5,000 requests per second under sustained traffic. LiteLLM is Python-based; Python overhead can add hundreds of microseconds at high concurrency. For low-volume indie-dev workflows the difference is invisible. For latency-critical production at scale, Bifrost performs measurably better. Bifrost is out of our catalog; from our catalog LiteLLM is the OSS proxy pick.

Can I migrate from a closed-source SaaS to OSS self-host?

Yes for routing logic. Application code calling OpenAI SDK works unchanged behind LiteLLM proxy. Cost-tracking and observability data does not migrate; you start fresh on the new tool. Plan one to three weeks for a non-trivial migration of routing logic plus dashboards. The honest framework: do not start a migration without budgeting design time; OSS self-host has different operational characteristics than managed SaaS, and direct ports rarely match feature for feature.

Why are LiteLLM, Helicone, and Langfuse all Y Combinator W23?

The LLM gateway category emerged in late 2022 and early 2023 as developers building on OpenAI hit the same problems: provider lock-in, no cost visibility, no caching, no fallbacks. Y Combinator funded several teams that winter (W23) at different layers: LiteLLM (OSS proxy), Helicone (observability-first proxy), Langfuse (OSS observability). The timing aligned with ChatGPT launch (Nov 2022) and the GPT-4 API release (March 2023).

EU data residency: which OSS picks store gateway data in the EU?

All three OSS picks self-host gives full control of where data lives, so EU residency is user-controlled on self-host. Langfuse Cloud is Berlin-based with default EU residency. LiteLLM Cloud has multi-region with EU option. Portkey Cloud has multi-region with EU on Pro and Enterprise. For EU-resident OSS use, all three self-host paths qualify; managed Cloud picks all offer EU regions on paid tiers.

How often is this guide updated?

We re-review pricing and feature changes annually at minimum, with mid-year refreshes when major vendor announcements happen. Portkey open-sourced under Apache 2.0 in March 2026 (a major shift). LiteLLM crossed 100 providers in 2024. Langfuse Core launched at twenty-nine dollars monthly in 2024. The lastReviewed date reflects the most recent editorial pass.

Subrupt Editorial

The team behind subrupt.com. We track subscriptions, surface cheaper alternatives, and publish buying guides where the score formula is on the page so you can recompute it yourself. We do not claim 30,000 hours of testing. What we claim is live pricing from our database, a transparent composite score, and honest savings math against a category baseline.

Last reviewed

Citations

Affiliate disclosure: Subrupt earns a commission when you switch to a service through our recommendation links. This never changes the price you pay. We only recommend services where there's a real cost or feature advantage for you, and our picks are based on the data on this page, not on which programs pay the most.

Related buying guides

Track your subscriptions on Subrupt

Add the Open Source LLM Gateway you pay for and see how much you'd save by switching.

Open dashboard

More buying guides

Independent rankings for the subscriptions worth paying for.

See all guides