<?xml version="1.0" encoding="UTF-8"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>CheapestInference | Blog</title><description/><link>https://docs.cheapestinference.com/</link><language>en</language><item><title>Qwen 3.5 vs GPT-5.4 vs Claude Opus 4.6 — same quality, fraction of the price</title><link>https://docs.cheapestinference.com/blog/qwen-3-5-vs-gpt-claude/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/qwen-3-5-vs-gpt-claude/</guid><pubDate>Thu, 26 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;You asked for this. After our &lt;a href=&quot;https://docs.cheapestinference.com/blog/open-source-models-are-production-ready/&quot;&gt;first benchmark post&lt;/a&gt;, the most requested model was Qwen 3.5. Here it is — &lt;strong&gt;4 models across 5 metrics&lt;/strong&gt;, same models in every chart:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Open-source:&lt;/strong&gt; Qwen3.5-397B-A17B (flagship), Qwen3.5-35B-A3B (efficient)
&lt;strong&gt;Proprietary:&lt;/strong&gt; GPT-5.4, Claude Opus 4.6&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;knowledge-mmlu-pro&quot;&gt;Knowledge: MMLU-Pro (%)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;88.5%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 397B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;87.8%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 35B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;85.3%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;82.0%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;GPT-5.4 leads at 88.5%, but Qwen3.5-397B is 0.7 points behind — statistically noise. The 35B with only 3B active parameters scores 85.3%, beating Opus by 3 points. The total spread across all four models is just 6.5 points.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Qwen3.5-397B matches GPT-5.4 at 5x less cost. The 35B beats Opus at 23x less.&lt;/strong&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;reasoning-gpqa-diamond&quot;&gt;Reasoning: GPQA Diamond (%)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;92.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;91.3%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 397B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;88.4%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 35B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;84.2%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Proprietary models lead on graduate-level reasoning. GPT-5.4 at 92% and Opus at 91.3% are strong. But Qwen3.5-397B at 88.4% is within 4 points — and costs $0.54/M vs $2.50 and $5.00. The 35B at 84.2% is still PhD-level performance for $0.22/M input.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;code-livecodebench-v6&quot;&gt;Code: LiveCodeBench v6 (%)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;84.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 397B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;83.6%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;76.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 35B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;74.6%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;The 397B essentially ties GPT-5.4 on competitive coding — 0.4 points apart. Both beat Opus by 8+ points. The 35B at 74.6% is within 2 points of Opus, at 1/23rd the price.&lt;/p&gt;
&lt;p&gt;For dedicated coding workloads, we also serve &lt;a href=&quot;https://docs.cheapestinference.com/pricing&quot;&gt;Qwen3-Coder-480B&lt;/a&gt; (SWE-bench Verified: 69.6%, comparable to Claude Sonnet 4).&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;speed-output-tokens-per-second&quot;&gt;Speed: output tokens per second&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 35B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;178 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 397B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;84 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~78 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;46 t/s&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;The 35B’s MoE architecture pays off — 178 tok/s is 2.3x faster than GPT-5.4 and 3.9x faster than Opus. Even the 397B flagship at 84 tok/s outpaces both proprietary models. This is what happens when only 3-17B parameters activate per token instead of the full model.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Speed data from &lt;a href=&quot;https://artificialanalysis.ai/leaderboards/models&quot;&gt;Artificial Analysis&lt;/a&gt;. Actual speeds on our infrastructure may differ.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;price-input-cost-per-million-tokens&quot;&gt;Price: input cost per million tokens&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 35B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;$0.22&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Qwen3.5 397B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;$0.54&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;$2.50&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;$5.00&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;This is the chart that matters. Opus costs &lt;strong&gt;23x more&lt;/strong&gt; than the 35B and &lt;strong&gt;9x more&lt;/strong&gt; than the 397B. GPT-5.4 costs &lt;strong&gt;5x more&lt;/strong&gt; than the 397B. The quality difference? Single-digit percentage points on every benchmark.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-full-picture&quot;&gt;The full picture&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
&lt;svg viewBox=&quot;-40 0 480 400&quot; xmlns=&quot;http://www.w3.org/2000/svg&quot;&gt;
  &lt;!-- Grid rings --&gt;
  &lt;polygon points=&quot;200,120 270,190 200,260 130,190&quot; fill=&quot;none&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/polygon&gt;
  &lt;polygon points=&quot;200,50 340,190 200,330 60,190&quot; fill=&quot;none&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Axes --&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;200&quot; y2=&quot;50&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;340&quot; y2=&quot;190&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;200&quot; y2=&quot;330&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;60&quot; y2=&quot;190&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;!-- GPT-5.4 — gray fill for reference --&gt;
  &lt;polygon points=&quot;200,57 312,190 200,312.5 145.4,190&quot; fill=&quot;#9A9490&quot; fill-opacity=&quot;0.08&quot; stroke=&quot;#9A9490&quot; stroke-width=&quot;2&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Claude Opus 4.6 — dashed --&gt;
  &lt;polygon points=&quot;200,113 305.4,190 200,236.6 167.8,190&quot; fill=&quot;none&quot; stroke=&quot;#6B6560&quot; stroke-width=&quot;1.5&quot; stroke-dasharray=&quot;6 3&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Qwen3.5-397B — indigo --&gt;
  &lt;polygon points=&quot;200,59.8 278.4,190 200,304.4 141.2,190&quot; fill=&quot;#6366F1&quot; fill-opacity=&quot;0.12&quot; stroke=&quot;#6366F1&quot; stroke-width=&quot;2.5&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Qwen3.5-35B — teal --&gt;
  &lt;polygon points=&quot;200,122.8 239.2,190 200,275.1 75.4,190&quot; fill=&quot;#14B8A6&quot; fill-opacity=&quot;0.1&quot; stroke=&quot;#14B8A6&quot; stroke-width=&quot;2&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Data points - 397B --&gt;
  &lt;circle cx=&quot;200&quot; cy=&quot;59.8&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;278.4&quot; cy=&quot;190&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;200&quot; cy=&quot;304.4&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;141.2&quot; cy=&quot;190&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;!-- Labels --&gt;
  &lt;text x=&quot;200&quot; y=&quot;30&quot; text-anchor=&quot;middle&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Code&lt;/text&gt;
  &lt;text x=&quot;355&quot; y=&quot;194&quot; text-anchor=&quot;start&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Reasoning&lt;/text&gt;
  &lt;text x=&quot;200&quot; y=&quot;355&quot; text-anchor=&quot;middle&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Knowledge&lt;/text&gt;
  &lt;text x=&quot;45&quot; y=&quot;194&quot; text-anchor=&quot;end&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Speed&lt;/text&gt;
  &lt;!-- Legend --&gt;
  &lt;rect x=&quot;40&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#6366F1&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;58&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;Qwen3.5 397B&lt;/text&gt;
  &lt;rect x=&quot;145&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#14B8A6&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;163&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;Qwen3.5 35B&lt;/text&gt;
  &lt;rect x=&quot;230&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#9A9490&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;248&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;GPT-5.4&lt;/text&gt;
  &lt;line x1=&quot;310&quot; y1=&quot;371&quot; x2=&quot;324&quot; y2=&quot;371&quot; stroke=&quot;#6B6560&quot; stroke-width=&quot;1.5&quot; stroke-dasharray=&quot;4 2&quot;&gt;&lt;/line&gt;
  &lt;text x=&quot;328&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;Opus 4.6&lt;/text&gt;
&lt;/svg&gt;
&lt;/div&gt;
&lt;p&gt;Quality only — no price axis. GPT-5.4 (gray) has the largest shape. Opus (dashed) is strong on reasoning and code. The 397B (indigo) nearly overlaps GPT-5.4 on code and knowledge. The 35B (teal) pulls hard left on speed — 178 tok/s is 2.3x faster than anything else here. Price tells its own story in the chart above.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;the-scorecard&quot;&gt;The scorecard&lt;/h2&gt;&lt;/div&gt;





















































&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Metric&lt;/th&gt;&lt;th&gt;Winner&lt;/th&gt;&lt;th&gt;Qwen3.5 397B&lt;/th&gt;&lt;th&gt;GPT-5.4&lt;/th&gt;&lt;th&gt;Claude Opus 4.6&lt;/th&gt;&lt;th&gt;Gap (397B vs best)&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Knowledge&lt;/strong&gt; (MMLU-Pro)&lt;/td&gt;&lt;td&gt;GPT-5.4&lt;/td&gt;&lt;td&gt;87.8%&lt;/td&gt;&lt;td&gt;88.5%&lt;/td&gt;&lt;td&gt;82.0%&lt;/td&gt;&lt;td&gt;-0.7 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Reasoning&lt;/strong&gt; (GPQA)&lt;/td&gt;&lt;td&gt;GPT-5.4&lt;/td&gt;&lt;td&gt;88.4%&lt;/td&gt;&lt;td&gt;92.0%&lt;/td&gt;&lt;td&gt;91.3%&lt;/td&gt;&lt;td&gt;-3.6 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Code&lt;/strong&gt; (LiveCodeBench)&lt;/td&gt;&lt;td&gt;GPT-5.4&lt;/td&gt;&lt;td&gt;83.6%&lt;/td&gt;&lt;td&gt;84.0%&lt;/td&gt;&lt;td&gt;76.0%&lt;/td&gt;&lt;td&gt;-0.4 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Speed&lt;/strong&gt; (tok/s)&lt;/td&gt;&lt;td&gt;Qwen3.5 397B&lt;/td&gt;&lt;td&gt;84 t/s&lt;/td&gt;&lt;td&gt;~78 t/s&lt;/td&gt;&lt;td&gt;46 t/s&lt;/td&gt;&lt;td&gt;1.1x faster&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Price&lt;/strong&gt; ($/M input)&lt;/td&gt;&lt;td&gt;Qwen3.5 397B&lt;/td&gt;&lt;td&gt;$0.54&lt;/td&gt;&lt;td&gt;$2.50&lt;/td&gt;&lt;td&gt;$5.00&lt;/td&gt;&lt;td&gt;4.6x cheaper&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;p&gt;&lt;strong&gt;Same weight class, different price tag.&lt;/strong&gt; The 397B trades 0.4–3.6 points on quality for 4.6x lower price and faster speed. It beats Opus on 4 out of 5 metrics outright.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Note: The Qwen3.5-35B-A3B ($0.22/M) scores 85.3% MMLU-Pro, 84.2% GPQA, 74.6% LiveCodeBench at 178 tok/s — beating Opus on knowledge and speed at 23x less cost. A different weight class, but worth considering if speed and price matter more than the last few quality points.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-real-question-what-are-you-paying-for&quot;&gt;The real question: what are you paying for?&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;The quality gap between Qwen3.5-397B and GPT-5.4 is &lt;strong&gt;0.7 points on knowledge, 0.4 points on code&lt;/strong&gt;. The price gap is &lt;strong&gt;4.6x&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Put it differently:&lt;/p&gt;






























&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Model&lt;/th&gt;&lt;th&gt;MMLU-Pro&lt;/th&gt;&lt;th&gt;Cost per quality point&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;Qwen3.5 35B&lt;/td&gt;&lt;td&gt;85.3%&lt;/td&gt;&lt;td&gt;$0.003 per point per M tokens&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Qwen3.5 397B&lt;/td&gt;&lt;td&gt;87.8%&lt;/td&gt;&lt;td&gt;$0.006 per point per M tokens&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;GPT-5.4&lt;/td&gt;&lt;td&gt;88.5%&lt;/td&gt;&lt;td&gt;$0.028 per point per M tokens&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Claude Opus 4.6&lt;/td&gt;&lt;td&gt;82.0%&lt;/td&gt;&lt;td&gt;$0.061 per point per M tokens&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;p&gt;Opus costs &lt;strong&gt;20x more per quality point&lt;/strong&gt; than the 35B — and scores lower. GPT-5.4 leads on quality but costs 5-10x more for single-digit advantages.&lt;/p&gt;
&lt;p&gt;For most workloads, the last 3% of benchmark performance isn’t worth a 5x price increase. And for workloads where it is — the 397B gets you within 1 point of GPT-5.4 at a fraction of the cost.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;also-available-specialized-qwen-models&quot;&gt;Also available: specialized Qwen models&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Beyond the general-purpose models, we serve two Qwen specialists:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Qwen3-Coder-480B&lt;/strong&gt; — SWE-bench Verified 69.6%, comparable to Claude Sonnet 4. Built for agentic coding.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Qwen3-235B-Thinking&lt;/strong&gt; — Chain-of-thought reasoning specialist. When you need the model to show its work.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Both available through the same API, same flat-rate plans.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;All Qwen 3.5 models are available now on our API. Flat rate from $20/mo, or pay-as-you-go credits. &lt;a href=&quot;https://cheapestinference.com/pricing&quot;&gt;See pricing and try it →&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Sources:&lt;/strong&gt; &lt;a href=&quot;https://huggingface.co/Qwen/Qwen3.5-397B-A17B&quot;&gt;Qwen3.5-397B Model Card&lt;/a&gt; · &lt;a href=&quot;https://huggingface.co/Qwen/Qwen3.5-35B-A3B&quot;&gt;Qwen3.5-35B Model Card&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/leaderboards/models&quot;&gt;Artificial Analysis Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/gpqa-diamond&quot;&gt;GPQA Diamond Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://openai.com/api/pricing/&quot;&gt;OpenAI Pricing&lt;/a&gt; · &lt;a href=&quot;https://platform.claude.com/docs/en/about-claude/pricing&quot;&gt;Anthropic Pricing&lt;/a&gt; · &lt;a href=&quot;https://livecodebench.github.io/leaderboard.html&quot;&gt;LiveCodeBench Leaderboard&lt;/a&gt;&lt;/p&gt;</content:encoded></item><item><title>OpenClaw is free. Running it is not.</title><link>https://docs.cheapestinference.com/blog/openclaw-cost-problem/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/openclaw-cost-problem/</guid><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;OpenClaw has 247,000 GitHub stars. It’s free, open-source, and runs locally. You install it, point it at an LLM, and it writes code, browses the web, queries databases, and executes files on your behalf.&lt;/p&gt;
&lt;p&gt;The agent is free. The inference is not.&lt;/p&gt;
&lt;p&gt;Every time OpenClaw calls a model, it re-sends the entire conversation history — every tool output, every file it read, every intermediate result. By iteration 20 of a typical task, the input context is 30,000+ tokens. By iteration 40, it’s past 100,000. And it sends this &lt;em&gt;every single request&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;This is not a bug. It’s how agents work. And it’s why running OpenClaw on pay-per-token APIs costs $300–600/month for active users — sometimes more.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;where-the-tokens-go&quot;&gt;Where the tokens go&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;We broke down token consumption for a typical OpenClaw coding task: “add authentication to an Express API.” The agent completed it in 38 tool calls.&lt;/p&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Context accumulation&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~280K tokens&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;System prompt (×38)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~156K tokens&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Tool outputs (files, etc.)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~70K tokens&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Agent output&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~19K tokens&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Total: &lt;strong&gt;~525,000 tokens for a single task&lt;/strong&gt;. The agent’s actual output — the code it wrote — was 19K tokens. The other 96% is overhead.&lt;/p&gt;
&lt;p&gt;On Claude Opus at $15/M input + $75/M output, that single task costs &lt;strong&gt;$9.18&lt;/strong&gt;. Run five tasks a day and you’re at &lt;strong&gt;$1,377/month&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;On DeepSeek V3.2 via a pay-per-token provider at $0.27/M input + $1.10/M output, the same task costs &lt;strong&gt;$0.16&lt;/strong&gt;. Better — but 20 tasks a day is still &lt;strong&gt;$96/month&lt;/strong&gt;, and that’s &lt;em&gt;one agent&lt;/em&gt;.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-three-cost-traps&quot;&gt;The three cost traps&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;We covered these in depth in &lt;a href=&quot;https://docs.cheapestinference.com/blog/why-your-ai-agent-needs-a-budget/&quot;&gt;Why your AI agent needs a budget&lt;/a&gt;, but here’s the OpenClaw-specific version:&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;1-context-grows-quadratically&quot;&gt;1. Context grows quadratically&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;OpenClaw reads files into context. If it reads a 2,000-token file at step 5, that file gets re-sent at steps 6, 7, 8… all the way to 38. That single file read costs 2,000 × 33 remaining steps = &lt;strong&gt;66,000 tokens&lt;/strong&gt; in re-transmission alone.&lt;/p&gt;
&lt;p&gt;Users report session contexts at 56–58% of the 400K context window during normal use. This isn’t a failure mode — it’s the architecture working as designed.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;2-system-prompt-is-a-fixed-tax&quot;&gt;2. System prompt is a fixed tax&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;OpenClaw’s system prompt is ~9,600 tokens. It gets sent with every request. Over 38 tool calls, that’s 365K tokens just in system prompts. You pay this whether the agent does useful work or not.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;3-wrong-model-for-the-job&quot;&gt;3. Wrong model for the job&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;OpenClaw defaults to a single model for everything. But not every tool call needs the same intelligence:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Reading a file and deciding what to edit? &lt;strong&gt;Llama 3.1 8B&lt;/strong&gt; handles this at 200 tokens/sec.&lt;/li&gt;
&lt;li&gt;Writing complex authentication logic? &lt;strong&gt;DeepSeek V3.2&lt;/strong&gt; or &lt;strong&gt;Kimi K2.5&lt;/strong&gt; is the right call.&lt;/li&gt;
&lt;li&gt;Formatting a config file? &lt;strong&gt;Any 8B model&lt;/strong&gt; is overkill but still cheaper than Opus.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;We wrote a full guide on this pattern: &lt;a href=&quot;https://docs.cheapestinference.com/blog/multi-model-architecture/&quot;&gt;Building a multi-model architecture&lt;/a&gt;. Routing agent requests to the right model can cut costs by 60–80% without reducing output quality.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-math-on-flat-rate-vs-pay-per-token&quot;&gt;The math on flat-rate vs. pay-per-token&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Here’s the comparison for an OpenClaw user running ~20 tasks/day:&lt;/p&gt;
&lt;div&gt;
  &lt;table&gt;
    &lt;tbody&gt;&lt;tr&gt;
      &lt;th&gt;Provider&lt;/th&gt;
      &lt;th&gt;Cost/task&lt;/th&gt;
      &lt;th&gt;20 tasks/day&lt;/th&gt;
      &lt;th&gt;Monthly&lt;/th&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;Claude Opus (direct)&lt;/td&gt;
      &lt;td&gt;$9.18&lt;/td&gt;
      &lt;td&gt;$183.60&lt;/td&gt;
      &lt;td&gt;$5,508&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;GPT-5.4 (direct)&lt;/td&gt;
      &lt;td&gt;$4.73&lt;/td&gt;
      &lt;td&gt;$94.60&lt;/td&gt;
      &lt;td&gt;$2,838&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;DeepSeek V3.2 (per-token)&lt;/td&gt;
      &lt;td&gt;$0.16&lt;/td&gt;
      &lt;td&gt;$3.20&lt;/td&gt;
      &lt;td&gt;$96&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;CheapestInference Pro&lt;/td&gt;
      &lt;td&gt;—&lt;/td&gt;
      &lt;td&gt;—&lt;/td&gt;
      &lt;td&gt;$50/mo flat&lt;/td&gt;
    &lt;/tr&gt;
  &lt;/tbody&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;p&gt;Flat-rate means you don’t care about context accumulation. The 280K tokens of context overhead that makes pay-per-token expensive? Irrelevant. The system prompt tax? Doesn’t matter. Your agent can call models 24/7 and the bill is the same.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;what-wed-actually-recommend&quot;&gt;What we’d actually recommend&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;If you’re running OpenClaw, here’s the setup we see working best:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;1. Use open-source models.&lt;/strong&gt; DeepSeek V3.2 and Kimi K2.5 score within 4 points of proprietary models on coding benchmarks (&lt;a href=&quot;https://docs.cheapestinference.com/blog/open-source-models-are-production-ready/&quot;&gt;the data&lt;/a&gt;). The gap doesn’t justify a 50x cost difference.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;2. Route by complexity.&lt;/strong&gt; Don’t send file reads and simple decisions to the same model as complex code generation. A router model costs fractions of a cent per classification. Full guide: &lt;a href=&quot;https://docs.cheapestinference.com/blog/multi-model-architecture/&quot;&gt;Multi-model architecture&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3. Set per-key budgets.&lt;/strong&gt; One API key per agent, each with a dollar-denominated budget that resets every few hours. When the budget runs out, the agent pauses instead of burning through your allocation. We built this into every key: &lt;a href=&quot;https://docs.cheapestinference.com/blog/why-your-ai-agent-needs-a-budget/&quot;&gt;Agent budgets explained&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;4. Handle rate limits automatically.&lt;/strong&gt; Budget caps mean your agent &lt;em&gt;will&lt;/em&gt; hit 429s. That’s the point — the cap is working. But OpenClaw kills the conversation when it gets a 429. The agent stops, and if you close the dashboard, that conversation is gone.&lt;/p&gt;
&lt;p&gt;We built an OpenClaw plugin that fixes this: &lt;a href=&quot;https://github.com/cheapestinference/openclaw-plugin-ratelimit-retry&quot;&gt;&lt;code dir=&quot;auto&quot;&gt;openclaw-ratelimit-retry&lt;/code&gt;&lt;/a&gt;. It hooks into &lt;code dir=&quot;auto&quot;&gt;agent_end&lt;/code&gt;, detects retriable 429s, parks the session on disk, and waits for the budget window to reset. Then it sends &lt;code dir=&quot;auto&quot;&gt;chat.send&lt;/code&gt; to the original session — resuming the conversation with its full transcript, as if you had typed a message.&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;span&gt;&lt;/span&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;openclaw&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;plugins&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;install&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;@cheapestinference/openclaw-ratelimit-retry&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;span&gt;~/.openclaw/config.yaml&lt;/span&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;plugins&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;  &lt;/span&gt;&lt;span&gt;ratelimit-retry&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;budgetWindowHours&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;5&lt;/span&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;# matches your CheapestInference budget reset&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;maxRetryAttempts&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;3&lt;/span&gt;&lt;span&gt;     &lt;/span&gt;&lt;span&gt;# give up after 3 consecutive 429s&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;checkIntervalMinutes&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;5&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;# check every 5 min for ready retries&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;The plugin is zero-dependency, persists across server restarts, deduplicates by session, and handles edge cases like sub-agents, queue overflow, and corrupted state files. If the retry itself hits a 429, it re-queues automatically. No tokens wasted on re-sending from scratch — the agent picks up exactly where it left off.&lt;/p&gt;
&lt;p&gt;This turns budget caps from “your agent crashes” into “your agent naps and wakes up.” Set it up once and forget about it.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;5. Consider flat-rate.&lt;/strong&gt; If your agent runs more than a few tasks per day, per-token pricing works against you. Every token of context overhead is money. On flat-rate, context overhead is free — use the full 128K window, re-send everything, let the agent work without constraint.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-irony&quot;&gt;The irony&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;OpenClaw is free because the code runs on your machine. But the valuable part — the intelligence — runs on someone else’s GPUs. The agent framework is the cheap part. Inference is the expensive part.&lt;/p&gt;
&lt;p&gt;Open-source models on flat-rate infrastructure flip this equation. The models are free. The inference is flat. The only variable cost left is your time.&lt;/p&gt;
&lt;p&gt;Point your OpenClaw &lt;code dir=&quot;auto&quot;&gt;base_url&lt;/code&gt; at &lt;code dir=&quot;auto&quot;&gt;https://api.cheapestinference.com/v1&lt;/code&gt; and find out what unconstrained agents actually cost: nothing more than you already budgeted.&lt;/p&gt;</content:encoded></item><item><title>Why your AI agent needs a budget</title><link>https://docs.cheapestinference.com/blog/why-your-ai-agent-needs-a-budget/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/why-your-ai-agent-needs-a-budget/</guid><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;There’s a pattern that plays out every week in AI Discord servers and GitHub issues: someone deploys an agent, goes to bed, and wakes up to a $400 bill from a loop that ran all night.&lt;/p&gt;
&lt;p&gt;Agents are not humans. They don’t get tired. They don’t notice when they’re repeating themselves. And they consume tokens at a rate that makes interactive chat look like a rounding error.&lt;/p&gt;
&lt;p&gt;If you’re running agents in production — or even in development — you need a budget. Here’s why, and how to implement one.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;agents-consume-1050x-more-tokens-than-humans&quot;&gt;Agents consume 10–50x more tokens than humans&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;A human chatting with an LLM sends a message, reads the response, thinks, types another message. Maybe 10 requests per hour, a few hundred tokens each.&lt;/p&gt;
&lt;p&gt;An agent running a tool loop does this:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;1. Read task description (system prompt + context)     → 4,000 tokens input&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;2. Call tool #1                                         → 500 tokens output&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;3. Receive tool result, re-send full context + result   → 5,200 tokens input&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;4. Call tool #2                                         → 500 tokens output&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;5. Receive result, re-send everything                   → 6,800 tokens input&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;6. ... repeat 20-40 times ...&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;Each iteration re-sends the entire conversation history. By step 20, the input context is 30,000+ tokens — and the agent sends it &lt;em&gt;every single time&lt;/em&gt;. A 40-step agent loop can consume 500,000+ tokens in a single task. That’s what a human user consumes in a week.&lt;/p&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Agent (40-step loop)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~500K tokens&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Agent (10-step loop)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~100K tokens&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Human (1 hour chat)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~10K tokens&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;This is normal behavior. The agent is doing its job. The problem is when it does its job &lt;em&gt;wrong&lt;/em&gt; — and nobody is watching.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-three-failure-modes-that-drain-budgets&quot;&gt;The three failure modes that drain budgets&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;&lt;h3 id=&quot;1-infinite-tool-loops&quot;&gt;1. Infinite tool loops&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;The agent calls a tool, gets an error, retries the same call, gets the same error, retries again. Without a loop detector or retry cap, this continues until your rate limit or budget hits zero.&lt;/p&gt;
&lt;p&gt;This is the most common failure mode. It happens when:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;An API the agent calls is temporarily down&lt;/li&gt;
&lt;li&gt;The agent’s output doesn’t match the tool’s expected input format&lt;/li&gt;
&lt;li&gt;The agent misinterprets the tool result and keeps “trying harder”&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;A single infinite loop can consume millions of tokens in minutes.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;2-context-accumulation&quot;&gt;2. Context accumulation&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;Every tool result gets appended to the conversation. The agent never summarizes or trims. By step 30, the input payload is 40K+ tokens, and most of it is irrelevant tool outputs from step 3.&lt;/p&gt;
&lt;p&gt;This isn’t a bug — it’s the default behavior of most agent frameworks. The context grows linearly with each step, and each step costs more than the last because the full context is re-sent.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;3-wrong-model-for-the-job&quot;&gt;3. Wrong model for the job&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;An agent using DeepSeek R1 (a reasoning model at ~30 tokens/second) for tasks that don’t require reasoning — file listing, simple classification, template generation — is burning expensive compute for no quality gain. R1 also produces internal chain-of-thought tokens that you pay for but never see.&lt;/p&gt;
&lt;p&gt;The fix is model routing — covered in our &lt;a href=&quot;https://docs.cheapestinference.com/blog/multi-model-architecture&quot;&gt;multi-model architecture guide&lt;/a&gt;. But even with routing, you need a budget as a backstop.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;what-happens-without-a-budget&quot;&gt;What happens without a budget&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Without a spending cap, any of these failures means:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Pay-as-you-go API&lt;/strong&gt;: The bill grows until you notice. Stories of $500+ surprise bills are common on forums. The provider has no reason to stop you — they’re selling tokens.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Self-hosted inference&lt;/strong&gt;: The agent consumes your entire GPU allocation, starving other workloads.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Shared platform&lt;/strong&gt;: One user’s agent consumes capacity that other users need.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;In all three cases, the damage scales with time. An agent that runs for 8 hours unattended can do 8 hours of damage.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;how-budget-caps-work&quot;&gt;How budget caps work&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;A budget cap is a dollar ceiling on how much a single key can spend in a time window. When the cap is reached, requests return a &lt;code dir=&quot;auto&quot;&gt;429 Too Many Requests&lt;/code&gt; error. No overage charges. No surprise bills. The agent stops, and you investigate.&lt;/p&gt;
&lt;p&gt;The key properties of a good budget system:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;1. Dollar-denominated, not token-denominated.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Token limits sound intuitive but don’t work across models. 100,000 tokens of Llama 3.1 8B costs $0.002. The same tokens on a large reasoning model costs 100x more. A dollar budget normalizes across all models automatically.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;2. Time-windowed with automatic reset.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;A budget that resets every few hours (e.g. every 5 hours) means a failure in one window doesn’t affect the next. The agent recovers automatically. If you set a one-time budget that never resets, you have to manually intervene every time the agent exhausts it.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3. Per-key, not per-account.&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;If you run 5 agents, each should have its own key and its own budget. One runaway agent should not starve the other four. Per-key budgets provide isolation — the same way containers isolate processes.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;designing-agents-that-handle-budget-limits-gracefully&quot;&gt;Designing agents that handle budget limits gracefully&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;A well-built agent treats a budget limit the same way a well-built web app treats a rate limit — as a normal operational condition, not an unexpected error.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;catch-429s-and-degrade&quot;&gt;Catch 429s and degrade&lt;/h3&gt;&lt;/div&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;from&lt;/span&gt;&lt;span&gt; openai &lt;/span&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; OpenAI, RateLimitError&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;client &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;OpenAI&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;base_url&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;https://api.cheapestinference.com/v1&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;api_key&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;sk_your_agent_key&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;agent_step&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;list&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt; -&gt; &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;try&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; response.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].message.content&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;except&lt;/span&gt;&lt;span&gt; RateLimitError:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;# Budget exhausted — save state, wait for reset&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;save_agent_state&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;[BUDGET_LIMIT] Agent paused. Will resume on next window.&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;div&gt;&lt;h3 id=&quot;monitor-spend-proactively&quot;&gt;Monitor spend proactively&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;Don’t wait for the 429. Check your remaining budget periodically and adjust behavior:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; requests&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;check_budget&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;api_key&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt; -&gt; &lt;/span&gt;&lt;span&gt;dict&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;span&gt;Check remaining budget via the usage endpoint.&lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;resp &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; requests.&lt;/span&gt;&lt;span&gt;get&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;https://api.cheapestinference.com/v1/usage&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;headers&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Authorization&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;f&lt;/span&gt;&lt;span&gt;&quot;Bearer &lt;/span&gt;&lt;span&gt;{api_key}&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; resp.&lt;/span&gt;&lt;span&gt;json&lt;/span&gt;&lt;span&gt;()&lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;budget&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;budget &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;check_budget&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;sk_your_agent_key&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;remaining &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; budget[&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;limit&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;] &lt;/span&gt;&lt;span&gt;-&lt;/span&gt;&lt;span&gt; budget[&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;spent&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; remaining &lt;/span&gt;&lt;span&gt;&amp;#x3C;&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0.01&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;# Less than $0.01 left — switch to cheapest model or pause&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;switch_to_model&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;meta-llama/llama-3.1-8b-instruct&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;div&gt;&lt;h3 id=&quot;set-retry-caps-in-your-agent-framework&quot;&gt;Set retry caps in your agent framework&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;Every agent framework has a way to limit retries. Use it:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;# LangChain&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;agent &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;create_react_agent&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;llm&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;llm&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;tools&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;tools&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;max_iterations&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;25&lt;/span&gt;&lt;span&gt;  &lt;/span&gt;&lt;span&gt;# Hard cap on tool loop iterations&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;# CrewAI&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;agent &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;Agent&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;researcher&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;max_iter&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;15&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;span&gt;  &lt;/span&gt;&lt;span&gt;# Maximum iterations per task&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;llm&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;llm&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;# Custom loop&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;MAX_STEPS&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;30&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;for&lt;/span&gt;&lt;span&gt; step &lt;/span&gt;&lt;span&gt;in&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;range&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;MAX_STEPS&lt;/span&gt;&lt;span&gt;):&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;result &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;agent_step&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;is_done&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;result&lt;/span&gt;&lt;span&gt;):&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;break&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;else&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;log.&lt;/span&gt;&lt;span&gt;warning&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Agent hit max steps without completing task&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;A max iteration cap is your first line of defense. The budget cap is your second.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;subscriptions-as-a-natural-budget-mechanism&quot;&gt;Subscriptions as a natural budget mechanism&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Pay-per-token pricing gives agents an open-ended credit line. Subscriptions invert this — you decide upfront how much to spend, and the platform enforces it.&lt;/p&gt;
&lt;p&gt;With a subscription plan on cheapestinference:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Each key gets a &lt;strong&gt;dollar budget that resets every 5 hours&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;When budget runs out → &lt;code dir=&quot;auto&quot;&gt;429&lt;/code&gt;, never overage charges&lt;/li&gt;
&lt;li&gt;You create &lt;strong&gt;unlimited keys&lt;/strong&gt; — one per agent, each with its own budget&lt;/li&gt;
&lt;li&gt;When your subscription expires, &lt;strong&gt;all keys are automatically revoked&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This means your worst case is bounded. A runaway agent burns through one 5-hour budget window and stops. It doesn’t burn through your monthly allocation, because the next window starts fresh with a new budget.&lt;/p&gt;
&lt;p&gt;For teams running multiple agents, the per-key isolation matters. Your research agent, your coding agent, and your monitoring agent each have independent budgets. If the research agent enters a loop, the others keep working.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-budget-stack-defense-in-depth&quot;&gt;The budget stack: defense in depth&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;No single mechanism catches every failure. Stack them:&lt;/p&gt;






























&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Layer&lt;/th&gt;&lt;th&gt;What it catches&lt;/th&gt;&lt;th&gt;When it triggers&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Max iterations&lt;/strong&gt; (code)&lt;/td&gt;&lt;td&gt;Runaway tool loops&lt;/td&gt;&lt;td&gt;After N steps&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Retry cap&lt;/strong&gt; (code)&lt;/td&gt;&lt;td&gt;Repeated failed calls&lt;/td&gt;&lt;td&gt;After N consecutive errors&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Budget cap&lt;/strong&gt; (platform)&lt;/td&gt;&lt;td&gt;All spending, any cause&lt;/td&gt;&lt;td&gt;When dollar limit is reached&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Subscription expiry&lt;/strong&gt; (platform)&lt;/td&gt;&lt;td&gt;Abandoned agents&lt;/td&gt;&lt;td&gt;When subscription period ends&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;p&gt;The first two are your responsibility as the developer. The last two are the platform’s. Together, they ensure that even if your code has a bug you haven’t found yet, the damage is capped.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;what-a-budgeted-agent-looks-like-in-practice&quot;&gt;What a budgeted agent looks like in practice&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Here’s a complete pattern for a production agent:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;from&lt;/span&gt;&lt;span&gt; openai &lt;/span&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; OpenAI, RateLimitError&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; requests&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; time&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;client &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;OpenAI&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;base_url&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;https://api.cheapestinference.com/v1&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;api_key&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;sk_agent_research&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;MAX_STEPS&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;30&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;BUDGET_WARN_THRESHOLD&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0.02&lt;/span&gt;&lt;span&gt;  &lt;/span&gt;&lt;span&gt;# Switch models when &amp;#x3C; $0.02 left&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;RETRY_LIMIT&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;3&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;run_agent&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;task&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;messages &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;system&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;You are a research agent. ...&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;},&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;user&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: task}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;model &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;consecutive_errors &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;for&lt;/span&gt;&lt;span&gt; step &lt;/span&gt;&lt;span&gt;in&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;range&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;MAX_STEPS&lt;/span&gt;&lt;span&gt;):&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;# Check budget every 5 steps&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; step &lt;/span&gt;&lt;span&gt;%&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;5&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;==&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;and&lt;/span&gt;&lt;span&gt; step &lt;/span&gt;&lt;span&gt;&gt;&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;budget &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;check_budget&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;sk_agent_research&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;remaining &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; budget[&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;limit&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;] &lt;/span&gt;&lt;span&gt;-&lt;/span&gt;&lt;span&gt; budget[&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;spent&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; remaining &lt;/span&gt;&lt;span&gt;&amp;#x3C;&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;BUDGET_WARN_THRESHOLD&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                &lt;/span&gt;&lt;/span&gt;&lt;span&gt;model &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;meta-llama/llama-3.1-8b-instruct&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;try&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;consecutive_errors &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;content &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; response.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].message.content&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;messages.&lt;/span&gt;&lt;span&gt;append&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;assistant&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: content}&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;is_task_complete&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;):&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; content&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;except&lt;/span&gt;&lt;span&gt; RateLimitError:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;save_agent_state&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;span&gt; step&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;f&lt;/span&gt;&lt;span&gt;&quot;Budget limit reached at step &lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;step&lt;/span&gt;&lt;span&gt;}&lt;/span&gt;&lt;span&gt;. State saved.&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;except&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;Exception&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;as&lt;/span&gt;&lt;span&gt; e:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;consecutive_errors &lt;/span&gt;&lt;span&gt;+=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;1&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; consecutive_errors &lt;/span&gt;&lt;span&gt;&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;RETRY_LIMIT&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;f&lt;/span&gt;&lt;span&gt;&quot;Aborting after &lt;/span&gt;&lt;span&gt;{RETRY_LIMIT}&lt;/span&gt;&lt;span&gt; consecutive errors: &lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;e&lt;/span&gt;&lt;span&gt;}&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Max steps reached. Partial results saved.&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;Three layers of protection:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Max 30 steps&lt;/strong&gt; — prevents infinite loops&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;3 consecutive error retry cap&lt;/strong&gt; — prevents retry storms&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Budget check every 5 steps&lt;/strong&gt; — degrades to cheaper model before hitting the hard cap&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;If all three fail, the platform’s budget cap catches it anyway.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-bottom-line&quot;&gt;The bottom line&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Running an AI agent without a budget is like running a process without memory limits — it works fine until it doesn’t, and then the damage is proportional to how long nobody noticed.&lt;/p&gt;
&lt;p&gt;Budget caps don’t limit what your agent can do. They limit what it can do &lt;em&gt;wrong&lt;/em&gt;. A properly budgeted agent completes the same tasks — it just can’t bankrupt you in the process.&lt;/p&gt;
&lt;p&gt;Set a budget. Set a retry cap. Set a max iteration count. Then let your agent run.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;We serve 70+ open-source models with per-key budget caps that reset every 5 hours. One subscription, unlimited keys, and the guarantee that a bad loop never turns into a bad bill. &lt;a href=&quot;https://cheapestinference.com/register&quot;&gt;Get started&lt;/a&gt; or &lt;a href=&quot;https://cheapestinference.com/platforms&quot;&gt;see how per-key plans work&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;</content:encoded></item><item><title>Building a multi-model architecture: route requests to the right LLM</title><link>https://docs.cheapestinference.com/blog/multi-model-architecture/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/multi-model-architecture/</guid><pubDate>Thu, 19 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;Using one model for everything is the simplest architecture. It’s also the most wasteful. A 685B-parameter reasoning model answering “what’s the weather?” is like hiring a PhD to sort mail.&lt;/p&gt;
&lt;p&gt;This guide covers how to use a small, fast model to classify incoming requests and route them to the right specialist. The result: lower latency, lower cost, and often better quality — because each model handles what it’s actually good at.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-problem-with-single-model-architectures&quot;&gt;The problem with single-model architectures&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Most applications start with one model:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;User request --&gt; Large Model --&gt; Response&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;This works, but every request — simple or complex — pays the same latency and cost penalty. When 60% of your traffic is simple classification, FAQ, or extraction, you’re burning expensive compute on tasks a small model handles equally well.&lt;/p&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Llama 3.1 8B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~200 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~60 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~30 t/s&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;The gap between Llama 8B and R1 is nearly 7x in throughput. Routing simple requests to the small model saves that difference on every request.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-multi-model-architecture&quot;&gt;The multi-model architecture&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;User request --&gt; Router (Llama 8B) --&gt; classify intent&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                                          &lt;/span&gt;&lt;/span&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                  &lt;/span&gt;&lt;/span&gt;&lt;span&gt;+-----------+-----------+-----------+&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                  &lt;/span&gt;&lt;/span&gt;&lt;span&gt;|           |           |           |&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;               &lt;/span&gt;&lt;/span&gt;&lt;span&gt;simple      general    reasoning     code&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                  &lt;/span&gt;&lt;/span&gt;&lt;span&gt;|           |           |           |&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;Llama 3.1 8B  DeepSeek   DeepSeek R1   Qwen3&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;V3.2                    Coder&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                  &lt;/span&gt;&lt;/span&gt;&lt;span&gt;|           |           |           |&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                  &lt;/span&gt;&lt;/span&gt;&lt;span&gt;+-----+-----+-----+-----+&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                     &lt;/span&gt;&lt;/span&gt;&lt;span&gt;Response&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;Two stages:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Classify&lt;/strong&gt; — The router model reads the user’s message and outputs a category. This takes ~0.2 seconds with Llama 8B.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Route&lt;/strong&gt; — Based on the category, forward the request to the appropriate specialist model.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The router adds minimal overhead (~200ms) but saves significant compute by keeping simple requests away from expensive models.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;step-1-classify-with-llama-31-8b&quot;&gt;Step 1: Classify with Llama 3.1 8B&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Llama 3.1 8B is the router. At ~200 t/s output speed, ~0.2s TTFT, and $0.02/M input tokens, the classification step costs almost nothing and completes before the user notices.&lt;/p&gt;
&lt;p&gt;The classification prompt is simple — you want a single-word category, not a conversation:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;from&lt;/span&gt;&lt;span&gt; openai &lt;/span&gt;&lt;span&gt;import&lt;/span&gt;&lt;span&gt; OpenAI&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;client &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;OpenAI&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;base_url&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;https://api.cheapestinference.com/v1&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;api_key&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;your-api-key&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;classify_request&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;user_message&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt; -&gt; &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;span&gt;Classify a user message into a routing category.&lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;meta-llama/llama-3.1-8b-instruct&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;system&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: (&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Classify the user&apos;s message into exactly one category. &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Respond with only the category name, nothing else.&lt;/span&gt;&lt;span&gt;\n\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;Categories:&lt;/span&gt;&lt;span&gt;\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;- simple: greetings, FAQ, simple factual questions&lt;/span&gt;&lt;span&gt;\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;- general: complex questions, analysis, writing, summarization&lt;/span&gt;&lt;span&gt;\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;- reasoning: math, logic, multi-step problems, science&lt;/span&gt;&lt;span&gt;\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;- code: code generation, debugging, refactoring, technical implementation&lt;/span&gt;&lt;span&gt;\n&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;                    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;- agent: tasks requiring tool use, web search, or multi-step execution&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                &lt;/span&gt;&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;},&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;user&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: user_message}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;],&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;max_tokens&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;10&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;temperature&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;category &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; response.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].message.content.&lt;/span&gt;&lt;span&gt;strip&lt;/span&gt;&lt;span&gt;().&lt;/span&gt;&lt;span&gt;lower&lt;/span&gt;&lt;span&gt;()&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;# Default to general if classification is unclear&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;valid &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; {&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;simple&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;general&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;reasoning&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;code&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;agent&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; category &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; category &lt;/span&gt;&lt;span&gt;in&lt;/span&gt;&lt;span&gt; valid &lt;/span&gt;&lt;span&gt;else&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;general&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;The key details: &lt;code dir=&quot;auto&quot;&gt;max_tokens=10&lt;/code&gt; because we only need one word. &lt;code dir=&quot;auto&quot;&gt;temperature=0&lt;/code&gt; for deterministic routing. The system prompt is explicit about format — no preamble, just the category.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;step-2-route-to-the-specialist&quot;&gt;Step 2: Route to the specialist&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Each category maps to a model optimized for that task:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;# Model routing table&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;ROUTE_TABLE&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; {&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;simple&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;meta-llama/llama-3.1-8b-instruct&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;general&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:   &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;reasoning&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-reasoner&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;code&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:      &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;qwen/qwen3-coder&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;agent&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:     &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;moonshotai/kimi-k2-5&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;route_request&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;user_message&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;conversation_history&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;list&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt; -&gt; &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;span&gt;Classify and route a request to the appropriate model.&lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;category &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;classify_request&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;user_message&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;model &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;ROUTE_TABLE&lt;/span&gt;&lt;span&gt;[category]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;conversation_history &lt;/span&gt;&lt;span&gt;+&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;user&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: user_message}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;],&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;stream&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;True&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;# Stream the response back&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;full_response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;for&lt;/span&gt;&lt;span&gt; chunk &lt;/span&gt;&lt;span&gt;in&lt;/span&gt;&lt;span&gt; response:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; chunk.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].delta.content:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;content &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; chunk.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].delta.content&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;            &lt;/span&gt;&lt;/span&gt;&lt;span&gt;full_response &lt;/span&gt;&lt;span&gt;+=&lt;/span&gt;&lt;span&gt; content&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;print&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;end&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;&quot;&quot;&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;flush&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;True&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; full_response&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;Notice that simple requests route back to Llama 8B — the same model that did the classification. For simple queries, the router overhead is effectively zero because the specialist is the same model and can reuse the warm connection.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;step-3-handle-edge-cases&quot;&gt;Step 3: Handle edge cases&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;The basic router works for most traffic, but production systems need a few refinements:&lt;/p&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;def&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;route_request_production&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;user_message&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;conversation_history&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;list&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;force_model&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;None&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;)&lt;/span&gt;&lt;span&gt; -&gt; tuple[&lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;str&lt;/span&gt;&lt;span&gt;]:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;span&gt;Production router with overrides and fallback.&lt;/span&gt;&lt;span&gt;&quot;&quot;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;# Allow explicit model override (for power users or testing)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;if&lt;/span&gt;&lt;span&gt; force_model:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;model &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; force_model&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;category &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;override&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;else&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;category &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;classify_request&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;span&gt;user_message&lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;model &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;ROUTE_TABLE&lt;/span&gt;&lt;span&gt;[category]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;try&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;conversation_history &lt;/span&gt;&lt;span&gt;+&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;user&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: user_message}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; response.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].message.content, category&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;
&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;except&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;Exception&lt;/span&gt;&lt;span&gt;:&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;# Fallback to V3.2 if the specialist is unavailable&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;fallback &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;        &lt;/span&gt;&lt;/span&gt;&lt;span&gt;response &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; client.chat.completions.&lt;/span&gt;&lt;span&gt;create&lt;/span&gt;&lt;span&gt;(&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;model&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;fallback&lt;/span&gt;&lt;span&gt;,&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;messages&lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt;conversation_history &lt;/span&gt;&lt;span&gt;+&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;[&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;                &lt;/span&gt;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;role&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;user&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;, &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;content&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: user_message}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;            &lt;/span&gt;&lt;span&gt;]&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;        &lt;/span&gt;&lt;span&gt;return&lt;/span&gt;&lt;span&gt; response.choices[&lt;/span&gt;&lt;span&gt;0&lt;/span&gt;&lt;span&gt;].message.content, &lt;/span&gt;&lt;span&gt;f&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;{&lt;/span&gt;&lt;span&gt;category&lt;/span&gt;&lt;span&gt;}&lt;/span&gt;&lt;span&gt;-&gt;fallback&quot;&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;Three patterns worth noting:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Force model&lt;/strong&gt; — Let callers bypass routing when they know what they need.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Fallback&lt;/strong&gt; — If a specialist model is down, fall back to V3.2. It handles everything reasonably well.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Return the category&lt;/strong&gt; — Log which route each request takes. You’ll need this data to tune the system.&lt;/li&gt;
&lt;/ol&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;cost-and-latency-comparison&quot;&gt;Cost and latency comparison&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Consider a workload of 1,000 requests with this distribution: 600 simple, 300 general, 70 reasoning, 30 code. Average 500 input tokens, 200 output tokens per request.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;single-model-approach-everything-on-v32&quot;&gt;Single-model approach (everything on V3.2)&lt;/h3&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Avg latency&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~4.5s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;All 1000 reqs&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;V3.2 only&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Every request waits for V3.2’s ~1.2s TTFT plus generation time at ~60 t/s. Simple questions get the same treatment as complex analysis.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;multi-model-approach-routed&quot;&gt;Multi-model approach (routed)&lt;/h3&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Simple (600)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~1.2s (8B)&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;General (300)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~4.7s (V3.2)&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Reasoning (70)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~9.0s (R1)&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Code (30)&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~3.5s (Coder)&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;The weighted average latency drops to approximately &lt;strong&gt;2.7s&lt;/strong&gt; — a 40% reduction. The 600 simple requests finish in ~1.2s instead of ~4.5s. That’s a 3.7x improvement for the majority of your traffic.&lt;/p&gt;
&lt;p&gt;The 70 reasoning requests are &lt;em&gt;slower&lt;/em&gt; individually (~9s vs ~4.5s) because R1 generates chain-of-thought tokens. But the quality on those specific requests is significantly better — R1 scores 50.2% on HLE versus V3.2’s 39.3%.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;You get faster averages &lt;em&gt;and&lt;/em&gt; better quality on the hard tail.&lt;/strong&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;real-example-a-support-chatbot&quot;&gt;Real example: a support chatbot&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;A customer support chatbot receives three types of requests:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;FAQ&lt;/strong&gt; (60%) — “What are your business hours?” / “How do I reset my password?”&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Complex support&lt;/strong&gt; (30%) — “I was charged twice for order #12345, can you investigate?”&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Technical issues&lt;/strong&gt; (10%) — “Your API returns 500 when I send multipart form data with UTF-8 filenames”&lt;/li&gt;
&lt;/ol&gt;
&lt;div&gt;&lt;h3 id=&quot;without-routing&quot;&gt;Without routing&lt;/h3&gt;&lt;/div&gt;
&lt;p&gt;All requests go to DeepSeek V3.2. FAQs get correct answers but with unnecessary latency. Technical issues get decent answers but miss edge cases that a code-specialized model would catch.&lt;/p&gt;
&lt;div&gt;&lt;h3 id=&quot;with-routing&quot;&gt;With routing&lt;/h3&gt;&lt;/div&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;SUPPORT_ROUTES&lt;/span&gt;&lt;span&gt; &lt;/span&gt;&lt;span&gt;=&lt;/span&gt;&lt;span&gt; {&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;simple&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;meta-llama/llama-3.1-8b-instruct&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,  &lt;/span&gt;&lt;span&gt;# FAQ, greetings&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;general&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:   &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,     &lt;/span&gt;&lt;span&gt;# Complex support&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;reasoning&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;: &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;deepseek/deepseek-chat-v3-0324&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,     &lt;/span&gt;&lt;span&gt;# Investigations&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;code&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:      &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;qwen/qwen3-coder&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,                   &lt;/span&gt;&lt;span&gt;# Technical issues&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;    &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;agent&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;:     &lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;moonshotai/kimi-k2-5&lt;/span&gt;&lt;span&gt;&quot;&lt;/span&gt;&lt;span&gt;,               &lt;/span&gt;&lt;span&gt;# Multi-step resolution&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;}&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;p&gt;FAQs resolve in ~1 second via Llama 8B. Complex support issues get V3.2’s full analytical capability. Technical problems route to Qwen3 Coder, which understands the code context better. If a support issue requires looking up order data via API, it routes to Kimi K2.5 for tool-assisted resolution.&lt;/p&gt;
&lt;p&gt;The classification step adds ~200ms. For the 60% of requests that drop from ~4.5s to ~1.2s, that’s an invisible cost.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;when-not-to-use-multi-model-routing&quot;&gt;When NOT to use multi-model routing&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Routing adds complexity. Skip it when:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;All your requests are the same type.&lt;/strong&gt; If you’re building a code editor, just use Qwen3 Coder. No routing needed.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;You have fewer than 100 requests/day.&lt;/strong&gt; The cost savings don’t justify the engineering overhead at low volume.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Latency doesn’t matter.&lt;/strong&gt; For batch processing or async workloads, a single capable model is simpler.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Your classification accuracy is low.&lt;/strong&gt; If the router misclassifies frequently, you get worse results than a single good model. Test the classifier on real traffic before deploying.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The sweet spot is high-volume applications with diverse request types — chatbots, API gateways, developer tools, and customer-facing products where response time directly affects user experience.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;implementation-checklist&quot;&gt;Implementation checklist&lt;/h2&gt;&lt;/div&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Log your traffic.&lt;/strong&gt; Before building a router, understand your request distribution. What percentage is simple? Complex? Code?&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Start with two tiers.&lt;/strong&gt; Llama 8B for simple, V3.2 for everything else. Add specialists only when you have data showing they help.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Measure classification accuracy.&lt;/strong&gt; Sample 100 requests, manually label them, compare against the router’s output. Target &gt;90% accuracy.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Add fallback.&lt;/strong&gt; Every specialist route should fall back to V3.2 if the specialist is unavailable.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Monitor per-route metrics.&lt;/strong&gt; Track latency, cost, and quality per category. This tells you where to optimize next.&lt;/li&gt;
&lt;/ol&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;All models in this guide are available through a single OpenAI-compatible API with no configuration changes between models. If you’re building a platform that needs LLM access for your users, &lt;a href=&quot;https://cheapestinference.com/platforms&quot;&gt;see how per-key plans work&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Sources:&lt;/strong&gt; &lt;a href=&quot;https://artificialanalysis.ai/leaderboards/models&quot;&gt;Artificial Analysis Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/models/deepseek-v3-2&quot;&gt;DeepSeek V3.2&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/humanitys-last-exam&quot;&gt;HLE Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://kimi-k25.com/blog/kimi-k2-5-benchmark&quot;&gt;Kimi K2.5 Benchmarks&lt;/a&gt;&lt;/p&gt;</content:encoded></item><item><title>How to choose the right open-source model for your task</title><link>https://docs.cheapestinference.com/blog/choosing-the-right-open-source-model/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/choosing-the-right-open-source-model/</guid><pubDate>Thu, 19 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;Most teams default to the biggest model available and call it a day. That works — until latency spikes, costs climb, and you realize a 8B-parameter model would have handled 60% of your requests just fine.&lt;/p&gt;
&lt;p&gt;This guide maps common use cases to specific models, with real throughput numbers from our infrastructure. No theory — just which model to pick and why.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;quick-decision-table&quot;&gt;Quick decision table&lt;/h2&gt;&lt;/div&gt;




























































&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Use case&lt;/th&gt;&lt;th&gt;Model&lt;/th&gt;&lt;th&gt;Why&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;General chat / assistants&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;DeepSeek V3.2&lt;/td&gt;&lt;td&gt;Best all-rounder. 85% MMLU-Pro, 73% SWE-bench, 60 t/s.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Complex reasoning&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;DeepSeek R1&lt;/td&gt;&lt;td&gt;50.2% on Humanity’s Last Exam. Chain-of-thought built in.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Code generation&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;Qwen3 Coder&lt;/td&gt;&lt;td&gt;Purpose-built for code. Strong on completions, refactoring, and debugging.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Agentic workflows&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;Kimi K2.5&lt;/td&gt;&lt;td&gt;334 t/s output, native tool use, 50.2% HLE with tools. Built for agents.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Vision / multimodal&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;Llama 4 Scout&lt;/td&gt;&lt;td&gt;17 active experts, 109B params, native image understanding.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Fast classification&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;Llama 3.1 8B&lt;/td&gt;&lt;td&gt;~200 t/s, 0.2s TTFT. Small enough for routing, tagging, extraction.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;General (budget)&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;GLM 4.7 Flash&lt;/td&gt;&lt;td&gt;Fast inference, competitive quality. Good when V3.2 is overkill.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Long context chat&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;MiniMax M2.5&lt;/td&gt;&lt;td&gt;Native long-context support. Handles large documents well.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Large general + reasoning&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;Qwen3 235B&lt;/td&gt;&lt;td&gt;235B MoE. Strong across benchmarks when you need maximum capability.&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Embeddings&lt;/strong&gt;&lt;/td&gt;&lt;td&gt;BGE Large&lt;/td&gt;&lt;td&gt;MTEB-tested. Solid retrieval quality for RAG pipelines.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;general-chat-and-assistants&quot;&gt;General chat and assistants&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: DeepSeek V3.2&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;DeepSeek V3.2 is the default choice for most workloads. It scores 85% on MMLU-Pro (beating Claude Opus 4.6’s 82%), 73% on SWE-bench Verified, and runs at ~60 tokens/second on our infrastructure.&lt;/p&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;334 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Llama 3.1 8B&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~200 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~60 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~30 t/s&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Broad knowledge, instruction following, multilingual, structured output.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Tasks that need step-by-step reasoning chains (use R1) or sub-100ms latency (use Llama 8B).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; You need a reliable general-purpose model that handles most tasks without specialization.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;complex-reasoning&quot;&gt;Complex reasoning&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: DeepSeek R1&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;R1 is a reasoning-first model. It produces explicit chain-of-thought tokens before its final answer. On Humanity’s Last Exam — a benchmark designed to be unsolvable by current models — R1 scores 50.2%, beating GPT-5.4 (41.6%) and Claude Opus 4.6 (40%).&lt;/p&gt;
&lt;p&gt;The tradeoff is speed. At ~30 t/s, R1 is the slowest model in our lineup. That’s expected — it’s generating reasoning tokens that never appear in the final output.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Math, science, logic puzzles, multi-step problems, anything where “thinking” helps.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Simple Q&amp;#x26;A, classification, or latency-sensitive applications.
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; The task requires multi-step deduction. If a human would need to “think through it,” R1 will outperform faster models.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;code-generation&quot;&gt;Code generation&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: Qwen3 Coder&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Qwen3 Coder is purpose-built for software engineering tasks — code completion, refactoring, debugging, and generation across languages. It’s trained specifically on code-heavy data and optimized for developer workflows.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Code completion, bug fixing, refactoring, test generation, multi-file edits.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; General conversation or non-code tasks (use V3.2).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; Code quality matters more than general knowledge. For mixed code-and-chat workflows, V3.2 or Kimi K2.5 may be more versatile.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;agentic-workflows&quot;&gt;Agentic workflows&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: Kimi K2.5&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Kimi K2.5 was designed for agentic use. It has native tool-calling support, runs at 334 t/s (the fastest model we serve), and scores 50.2% on HLE when using tools — matching R1’s reasoning-only score.&lt;/p&gt;
&lt;p&gt;The speed matters for agents. Each tool call is a round trip: the model generates a function call, the tool executes, the result goes back to the model. At 334 t/s and 0.31s TTFT, Kimi completes multi-step agent loops in seconds where slower models take minutes.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Tool use, function calling, multi-step task execution, fast iteration loops.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Pure reasoning without tools (R1 is better). Code-only tasks (Qwen3 Coder is more specialized).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; Your application involves tool calling, API interactions, or multi-step agent orchestration where speed compounds.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;vision-and-multimodal&quot;&gt;Vision and multimodal&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: Llama 4 Scout&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Llama 4 Scout is Meta’s mixture-of-experts multimodal model — 109B total parameters with 17 active experts. It handles text and images natively, making it the pick for tasks that require visual understanding alongside language.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Image description, visual Q&amp;#x26;A, document understanding, chart interpretation.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Text-only tasks where you’re paying for vision capability you don’t use (use V3.2).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; Your input includes images. For text-only workloads, other models are more efficient.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;fast-classification-and-routing&quot;&gt;Fast classification and routing&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: Llama 3.1 8B&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;At 8 billion parameters, Llama 3.1 8B runs at ~200 t/s with approximately 0.2s time to first token. It’s the right choice for tasks where speed matters more than depth: intent classification, sentiment analysis, entity extraction, content filtering, and request routing.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Classification, tagging, extraction, routing decisions, simple Q&amp;#x26;A, content moderation.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Complex reasoning, long-form generation, or tasks requiring deep world knowledge.
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; You need results in under a second and the task is well-defined. Also ideal as the router model in a multi-model architecture.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;budget-general-use&quot;&gt;Budget general use&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: GLM 4.7 Flash&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;GLM 4.7 Flash delivers competitive quality at fast inference speeds. When DeepSeek V3.2 is more capability than you need — simple conversations, basic summarization, FAQ bots — GLM 4.7 Flash gets the job done efficiently.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Simple chat, summarization, translation, basic Q&amp;#x26;A.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Complex reasoning or tasks where benchmark-leading quality matters.
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; You want good-enough quality with better speed and lower cost than the largest models.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;long-context&quot;&gt;Long context&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: MiniMax M2.5&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;MiniMax M2.5 handles long context windows natively. For workloads that involve ingesting large documents, long conversation histories, or extensive codebases, M2.5 maintains coherence across the full context.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Document analysis, long conversations, large-context summarization.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Short, simple tasks where context length is irrelevant (use Llama 8B or GLM Flash).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; Your input regularly exceeds what smaller-context models handle well.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;maximum-capability&quot;&gt;Maximum capability&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: Qwen3 235B&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;Qwen3 235B is a large mixture-of-experts model that competes across the full benchmark spectrum. When you need the highest possible quality and latency is not the primary constraint, Qwen3 235B delivers.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Broad capability across reasoning, knowledge, and generation. Strong multilingual support.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Latency-sensitive applications (large model, slower inference).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; You need top-tier quality and can tolerate higher latency. Good for batch processing and offline tasks.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;embeddings&quot;&gt;Embeddings&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;&lt;strong&gt;Pick: BGE Large&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;BGE Large (BAAI General Embedding) is a well-tested embedding model for retrieval-augmented generation. It performs well on MTEB benchmarks and produces dense vectors suitable for semantic search, document retrieval, and clustering.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Good at:&lt;/strong&gt; Semantic search, RAG pipelines, document similarity, clustering.
&lt;strong&gt;Not ideal for:&lt;/strong&gt; Generative tasks (it’s an embedding model, not a chat model).
&lt;strong&gt;Pick over alternatives when:&lt;/strong&gt; You need vector embeddings for search or retrieval. Pair it with a generative model for the full RAG pipeline.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-decision-tree&quot;&gt;The decision tree&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;&lt;figure&gt;&lt;figcaption&gt;&lt;/figcaption&gt;&lt;pre&gt;&lt;code&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;What&apos;s your task?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need to understand images?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; Llama 4 Scout&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need step-by-step reasoning? (math, logic, science)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; DeepSeek R1 (~30 t/s, but highest reasoning quality)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need tool calling / agent loops?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; Kimi K2.5 (334 t/s, native tool use)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need code generation / editing?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; Qwen3 Coder (purpose-built for code)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need embeddings for search/RAG?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; BGE Large&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need sub-200ms response?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; Llama 3.1 8B (~200 t/s, 0.2s TTFT)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need long context (large documents)?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; MiniMax M2.5&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- Need maximum quality, latency flexible?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|   YES --&gt; Qwen3 235B&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;|&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;+-- General purpose, good balance?&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;div&gt;&lt;span&gt;&lt;span&gt;    &lt;/span&gt;&lt;/span&gt;&lt;span&gt;YES --&gt; DeepSeek V3.2 (default choice)&lt;/span&gt;&lt;/div&gt;&lt;/div&gt;&lt;/code&gt;&lt;/pre&gt;&lt;div&gt;&lt;div aria-live=&quot;polite&quot;&gt;&lt;/div&gt;&lt;/div&gt;&lt;/figure&gt;&lt;/div&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-8020-rule&quot;&gt;The 80/20 rule&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;You don’t need ten models to cover most workloads.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Llama 3.1 8B handles 60% of requests.&lt;/strong&gt; Classification, routing, simple Q&amp;#x26;A, extraction, content filtering. Fast and cheap.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;DeepSeek V3.2 handles 30%.&lt;/strong&gt; General chat, complex instructions, knowledge-intensive tasks. The reliable all-rounder.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Specialized models handle the last 10%.&lt;/strong&gt; R1 for hard reasoning. Kimi K2.5 for agent loops. Qwen3 Coder for code. BGE Large for embeddings.&lt;/p&gt;
&lt;p&gt;Start with Llama 8B + V3.2. Add specialists only when you have evidence that general models aren’t performing on specific task categories. Measure first, specialize second.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;All models are available through a single OpenAI-compatible API. If you’re building a platform that needs LLM access for your users, &lt;a href=&quot;https://cheapestinference.com/platforms&quot;&gt;see how per-key plans work&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Sources:&lt;/strong&gt; &lt;a href=&quot;https://artificialanalysis.ai/leaderboards/models&quot;&gt;Artificial Analysis Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://www.swebench.com/&quot;&gt;SWE-bench Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://kimi-k25.com/blog/kimi-k2-5-benchmark&quot;&gt;Kimi K2.5 Benchmarks&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/models/deepseek-v3-2&quot;&gt;DeepSeek V3.2&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/humanitys-last-exam&quot;&gt;HLE Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/mmlu-pro&quot;&gt;MMLU-Pro Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://huggingface.co/spaces/mteb/leaderboard&quot;&gt;MTEB Leaderboard&lt;/a&gt;&lt;/p&gt;</content:encoded></item><item><title>Open-source models are production-ready. Here&apos;s the proof.</title><link>https://docs.cheapestinference.com/blog/open-source-models-are-production-ready/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/open-source-models-are-production-ready/</guid><pubDate>Thu, 19 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;There’s a persistent assumption in the industry: open-source models are fine for experimentation, but production workloads need GPT-5 or Claude Opus. We run open-source models in production every day. Here’s what the benchmarks actually say.&lt;/p&gt;
&lt;p&gt;We’re comparing &lt;strong&gt;5 models across 5 metrics&lt;/strong&gt; — the same models in every chart, no cherry-picking:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Open-source (available via our API):&lt;/strong&gt; DeepSeek V3.2, DeepSeek R1, Kimi K2.5
&lt;strong&gt;Proprietary (reference):&lt;/strong&gt; Claude Opus 4.6, GPT-5.4&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;code-quality-swe-bench-verified--resolved&quot;&gt;Code quality: SWE-bench Verified (% resolved)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;80.8%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~80.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;76.8%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;73.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;57.6%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Proprietary models lead here. Opus 4.6 and GPT-5.4 are within a point of each other at ~80%. Kimi K2.5 is 4 points behind at 76.8% — competitive but not leading. R1 is a reasoning model, not optimized for code.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;reasoning-humanitys-last-exam&quot;&gt;Reasoning: Humanity’s Last Exam (%)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5 *&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;50.2%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;50.2%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;41.6%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;40.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;39.3%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Open-source wins decisively. R1 hits 50.2% and Kimi K2.5 matches it with tool-use enabled (*without tools: 31.5%). Both beat Opus 4.6 (40%) and GPT-5.4 (41.6%). V3.2 is roughly at Opus level — it’s a general model, not a reasoning specialist.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;*Kimi K2.5’s HLE score uses its agentic mode with tool access. This is how the model is designed to be used in production.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;knowledge-mmlu-pro&quot;&gt;Knowledge: MMLU-Pro (%)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;88.5%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;87.1%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;85.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;84.0%&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;82.0%&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;GPT-5.4 leads narrowly at 88.5%, but Kimi K2.5 is 1.4 points behind and all three open-source models beat Opus 4.6. The gap across all 5 models is only 6.5 points — this benchmark is nearly saturated.&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;speed-output-tokens-per-second&quot;&gt;Speed: output tokens per second&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;334 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~78 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~60 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;46 t/s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~30 t/s&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Kimi K2.5 at 334 tok/s is in a different league — 4x faster than GPT-5.4, 7x faster than Opus 4.6. R1 is the slowest (expected — it’s a reasoning model producing chain-of-thought tokens).&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;latency-time-to-first-token-seconds&quot;&gt;Latency: time to first token (seconds)&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
  &lt;div&gt;
    &lt;span&gt;Kimi K2.5&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;0.31s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;GPT-5.4&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~0.95s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek V3.2&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;1.18s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;DeepSeek R1&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;~2.0s&lt;/span&gt;
  &lt;/div&gt;
  &lt;div&gt;
    &lt;span&gt;Claude Opus 4.6&lt;/span&gt;
    &lt;div&gt;&lt;/div&gt;
    &lt;span&gt;2.48s&lt;/span&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;Lower is better. Kimi K2.5 responds 8x faster than Opus 4.6 and 3x faster than GPT-5.4. Even V3.2 beats both proprietary models. Opus 4.6 is the slowest model in this comparison.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Speed and TTFT measured on our production infrastructure. Claude and GPT-5.4 data from Artificial Analysis.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-full-picture&quot;&gt;The full picture&lt;/h2&gt;&lt;/div&gt;
&lt;div&gt;
&lt;svg viewBox=&quot;-80 0 560 410&quot; xmlns=&quot;http://www.w3.org/2000/svg&quot;&gt;
  &lt;!-- Grid lines --&gt;
  &lt;polygon points=&quot;200,120 266.6,168.4 241.1,246.6 158.9,246.6 133.4,168.4&quot; fill=&quot;none&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/polygon&gt;
  &lt;polygon points=&quot;200,50 333.1,146.7 282.3,303.3 117.7,303.3 66.9,146.7&quot; fill=&quot;none&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Axes --&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;200&quot; y2=&quot;50&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;333.1&quot; y2=&quot;146.7&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;282.3&quot; y2=&quot;303.3&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;117.7&quot; y2=&quot;303.3&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;line x1=&quot;200&quot; y1=&quot;190&quot; x2=&quot;66.9&quot; y2=&quot;146.7&quot; stroke=&quot;#E8E5DF&quot; stroke-width=&quot;1&quot;&gt;&lt;/line&gt;
  &lt;!-- Kimi K2.5 — indigo --&gt;
  &lt;polygon points=&quot;200,57 333.1,146.7 280.6,301 117.7,303.3 66.9,146.7&quot; fill=&quot;#6366F1&quot; fill-opacity=&quot;0.12&quot; stroke=&quot;#6366F1&quot; stroke-width=&quot;2.5&quot;&gt;&lt;/polygon&gt;
  &lt;!-- DeepSeek V3.2 — teal --&gt;
  &lt;polygon points=&quot;200,64 303.9,156.3 279,298.7 185.2,210.4 120.1,164&quot; fill=&quot;#14B8A6&quot; fill-opacity=&quot;0.08&quot; stroke=&quot;#14B8A6&quot; stroke-width=&quot;2&quot;&gt;&lt;/polygon&gt;
  &lt;!-- DeepSeek R1 — amber --&gt;
  &lt;polygon points=&quot;200,90.6 333.1,146.7 278.2,297.6 192.6,200.2 170.7,180.5&quot; fill=&quot;#F59E0B&quot; fill-opacity=&quot;0.08&quot; stroke=&quot;#F59E0B&quot; stroke-width=&quot;2&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Claude Opus 4.6 — gray --&gt;
  &lt;polygon points=&quot;200,50 306.5,155.4 276.5,295.3 188.5,205.9 200,190&quot; fill=&quot;none&quot; stroke=&quot;#9A9490&quot; stroke-width=&quot;2&quot;&gt;&lt;/polygon&gt;
  &lt;!-- GPT-5.4 — dark gray dashed --&gt;
  &lt;polygon points=&quot;200,51.4 310.5,154.1 282.3,303.3 181.1,216.1 105.5,159.3&quot; fill=&quot;none&quot; stroke=&quot;#6B6560&quot; stroke-width=&quot;1.5&quot; stroke-dasharray=&quot;6 3&quot;&gt;&lt;/polygon&gt;
  &lt;!-- Data points --&gt;
  &lt;circle cx=&quot;200&quot; cy=&quot;57&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;333.1&quot; cy=&quot;146.7&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;280.6&quot; cy=&quot;301&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;117.7&quot; cy=&quot;303.3&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;circle cx=&quot;66.9&quot; cy=&quot;146.7&quot; r=&quot;3.5&quot; fill=&quot;#6366F1&quot;&gt;&lt;/circle&gt;
  &lt;!-- Labels --&gt;
  &lt;text x=&quot;200&quot; y=&quot;30&quot; text-anchor=&quot;middle&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Code&lt;/text&gt;
  &lt;text x=&quot;345&quot; y=&quot;142&quot; text-anchor=&quot;start&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Reasoning&lt;/text&gt;
  &lt;text x=&quot;290&quot; y=&quot;325&quot; text-anchor=&quot;start&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Knowledge&lt;/text&gt;
  &lt;text x=&quot;110&quot; y=&quot;325&quot; text-anchor=&quot;end&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Speed&lt;/text&gt;
  &lt;text x=&quot;55&quot; y=&quot;142&quot; text-anchor=&quot;end&quot; font-size=&quot;13&quot; font-weight=&quot;600&quot; fill=&quot;#1A1A1A&quot;&gt;Latency&lt;/text&gt;
  &lt;!-- Legend row 1 --&gt;
  &lt;rect x=&quot;-10&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#6366F1&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;8&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;Kimi K2.5&lt;/text&gt;
  &lt;rect x=&quot;75&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#14B8A6&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;93&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;DeepSeek V3.2&lt;/text&gt;
  &lt;rect x=&quot;185&quot; y=&quot;370&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#F59E0B&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;203&quot; y=&quot;374&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;DeepSeek R1&lt;/text&gt;
  &lt;!-- Legend row 2 --&gt;
  &lt;rect x=&quot;-10&quot; y=&quot;386&quot; width=&quot;14&quot; height=&quot;3&quot; rx=&quot;1&quot; fill=&quot;#9A9490&quot;&gt;&lt;/rect&gt;
  &lt;text x=&quot;8&quot; y=&quot;390&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;Claude Opus 4.6&lt;/text&gt;
  &lt;line x1=&quot;110&quot; y1=&quot;387&quot; x2=&quot;124&quot; y2=&quot;387&quot; stroke=&quot;#6B6560&quot; stroke-width=&quot;1.5&quot; stroke-dasharray=&quot;4 2&quot;&gt;&lt;/line&gt;
  &lt;text x=&quot;128&quot; y=&quot;390&quot; font-size=&quot;9&quot; fill=&quot;#6B6560&quot;&gt;GPT-5.4&lt;/text&gt;
&lt;/svg&gt;
&lt;/div&gt;
&lt;div&gt;&lt;h2 id=&quot;the-scorecard&quot;&gt;The scorecard&lt;/h2&gt;&lt;/div&gt;















































&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Metric&lt;/th&gt;&lt;th&gt;Winner&lt;/th&gt;&lt;th&gt;Open-source&lt;/th&gt;&lt;th&gt;Proprietary&lt;/th&gt;&lt;th&gt;Gap&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Code&lt;/strong&gt; (SWE-bench)&lt;/td&gt;&lt;td&gt;Opus 4.6&lt;/td&gt;&lt;td&gt;Kimi 76.8%&lt;/td&gt;&lt;td&gt;Opus 80.8%&lt;/td&gt;&lt;td&gt;-4 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Reasoning&lt;/strong&gt; (HLE)&lt;/td&gt;&lt;td&gt;R1&lt;/td&gt;&lt;td&gt;R1 50.2%&lt;/td&gt;&lt;td&gt;GPT-5.4 41.6%&lt;/td&gt;&lt;td&gt;+8.6 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Knowledge&lt;/strong&gt; (MMLU-Pro)&lt;/td&gt;&lt;td&gt;GPT-5.4&lt;/td&gt;&lt;td&gt;Kimi 87.1%&lt;/td&gt;&lt;td&gt;GPT-5.4 88.5%&lt;/td&gt;&lt;td&gt;-1.4 pts&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Speed&lt;/strong&gt; (tok/s)&lt;/td&gt;&lt;td&gt;Kimi K2.5&lt;/td&gt;&lt;td&gt;334 t/s&lt;/td&gt;&lt;td&gt;GPT-5.4 78 t/s&lt;/td&gt;&lt;td&gt;4.3x faster&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;&lt;strong&gt;Latency&lt;/strong&gt; (TTFT)&lt;/td&gt;&lt;td&gt;Kimi K2.5&lt;/td&gt;&lt;td&gt;0.31s&lt;/td&gt;&lt;td&gt;GPT-5.4 0.95s&lt;/td&gt;&lt;td&gt;3x faster&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;p&gt;&lt;strong&gt;Open-source wins 3 out of 5.&lt;/strong&gt; Proprietary models lead on Code (by 4 points) and Knowledge (by 1.4 points). Open-source leads on Reasoning (by 8.6 points), Speed (by 4.3x), and Latency (by 3x).&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Note: Kimi K2.5’s HLE score (50.2%) uses tool-augmented mode. Without tools it scores 31.5%. DeepSeek R1’s 50.2% is pure chain-of-thought reasoning without tools.&lt;/em&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;what-production-ready-actually-means&quot;&gt;What “production-ready” actually means&lt;/h2&gt;&lt;/div&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Reliable enough.&lt;/strong&gt; Consistent quality across thousands of requests.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Fast enough.&lt;/strong&gt; Kimi K2.5 at 334 tok/s and 0.31s TTFT. That’s real-time.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Capable enough.&lt;/strong&gt; Within 4 points of the best proprietary model on code, ahead on reasoning.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Predictable.&lt;/strong&gt; Versioned models that don’t change without warning.&lt;/li&gt;
&lt;/ol&gt;
&lt;div&gt;&lt;h2 id=&quot;the-real-advantage-control&quot;&gt;The real advantage: control&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Proprietary models change under you. Fine one day, different behavior the next. No changelog, no warning. Open-source models are versioned — DeepSeek V3.2 behaves the same tomorrow as today. You choose when to upgrade.&lt;/p&gt;
&lt;p&gt;For production workloads, that predictability is worth more than a marginal quality edge on any single benchmark.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;We serve 70+ open-source models through a single API. If you’re building a platform that needs LLM access for your users, &lt;a href=&quot;https://cheapestinference.com/platforms&quot;&gt;see how per-key plans work&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Sources:&lt;/strong&gt; &lt;a href=&quot;https://artificialanalysis.ai/leaderboards/models&quot;&gt;Artificial Analysis Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://www.swebench.com/&quot;&gt;SWE-bench Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://kimi-k25.com/blog/kimi-k2-5-benchmark&quot;&gt;Kimi K2.5 Benchmarks&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/models/deepseek-v3-2&quot;&gt;DeepSeek V3.2&lt;/a&gt; · &lt;a href=&quot;https://openai.com/api/pricing/&quot;&gt;OpenAI Pricing&lt;/a&gt; · &lt;a href=&quot;https://platform.claude.com/docs/en/about-claude/pricing&quot;&gt;Anthropic Pricing&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/humanitys-last-exam&quot;&gt;HLE Leaderboard&lt;/a&gt; · &lt;a href=&quot;https://artificialanalysis.ai/evaluations/mmlu-pro&quot;&gt;MMLU-Pro Leaderboard&lt;/a&gt;&lt;/p&gt;</content:encoded></item><item><title>What it takes to build your own LLM inference platform</title><link>https://docs.cheapestinference.com/blog/build-your-own-inference-platform/</link><guid isPermaLink="true">https://docs.cheapestinference.com/blog/build-your-own-inference-platform/</guid><pubDate>Thu, 19 Mar 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;If you’re building a SaaS that needs to give users access to LLMs, you have two options: build the infrastructure yourself, or use a platform that does it for you. Here’s what “build it yourself” actually looks like.&lt;/p&gt;
&lt;p&gt;This isn’t theoretical. We built this. Here’s every component, what it does, and what alternatives exist.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;0-model-access--the-first-problem&quot;&gt;0. Model access — the first problem&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Before you write a single line of code, you need access to models.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Self-host on your own hardware&lt;/strong&gt;: Buy GPUs, rent datacenter space, run the models yourself. Full control, best unit economics at scale — but massive upfront cost and you’re limited to the models you can afford to deploy. Running DeepSeek V3.2 requires multiple high-end GPUs. Running 70+ models? You’d need a data center.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Rent infrastructure&lt;/strong&gt;: Use GPU clouds like Vast.ai, AWS, Hetzner, CoreWeave, or Lambda. No hardware to buy, but you still manage deployments, scaling, and failover. Costs add up fast — a single H100 runs $2-4/hr.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Use an inference provider&lt;/strong&gt;: Sign agreements with DeepInfra, Together.ai, Fireworks, etc. who already have the models deployed. Pay per token, no GPU management. But you depend on their availability, pricing, and terms. If they change prices or drop a model, you need a plan B.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Mix&lt;/strong&gt;: Most serious platforms end up here. Own hardware for high-volume models where the unit economics justify it, rented GPUs for burst capacity, and provider agreements for the long tail of models nobody runs enough to self-host.&lt;/p&gt;
&lt;p&gt;Self-hosting 70+ models on your own is economically unrealistic. The real question is where to draw the line between own infra, rented compute, and providers.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;1-serving-engine&quot;&gt;1. Serving engine&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;If you self-host or rent GPUs, you need software to serve the models:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;vLLM&lt;/strong&gt; — most popular, good throughput, active community&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;TGI&lt;/strong&gt; (Text Generation Inference) — Hugging Face’s solution, solid for single-model deployments&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;TensorRT-LLM&lt;/strong&gt; — NVIDIA’s optimized engine, best raw performance but harder to set up&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;SGLang&lt;/strong&gt; — newer, fast, good for structured generation&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;You’ll also need to handle model weights, quantization, scaling across GPUs, and failover when a node goes down. This is a full-time ops job.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;2-api-proxy-layer&quot;&gt;2. API proxy layer&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Your users shouldn’t hit the inference backend directly. You need a proxy that:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Translates between API formats (OpenAI, Anthropic)&lt;/li&gt;
&lt;li&gt;Routes requests to the right model/provider&lt;/li&gt;
&lt;li&gt;Injects authentication&lt;/li&gt;
&lt;li&gt;Handles retries and failover&lt;/li&gt;
&lt;li&gt;Strips provider headers so users don’t know your backend&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Options:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Build from scratch with Express/Fastify + http-proxy-middleware&lt;/li&gt;
&lt;li&gt;Use an open-source gateway: LiteLLM, Portkey, Kong AI Gateway, MLflow Gateway&lt;/li&gt;
&lt;li&gt;Use a managed gateway: Helicone, Braintrust, Promptlayer&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Each has trade-offs. Open-source gateways give you control but you manage the deployment. Managed gateways are easier but add latency and cost.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;3-authentication&quot;&gt;3. Authentication&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Two layers:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;User auth (dashboard login)&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Firebase Auth, Auth0, Clerk, Supabase Auth, or roll your own&lt;/li&gt;
&lt;li&gt;Supports email, Google, GitHub, wallet signatures&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;API key auth (inference requests)&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Generate API keys per user&lt;/li&gt;
&lt;li&gt;Validate on every request before proxying&lt;/li&gt;
&lt;li&gt;Store key metadata (plan, rate limits, owner)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is where it gets interesting for platforms. You need &lt;strong&gt;per-key plans&lt;/strong&gt; — each key with its own rate limits and usage tracking. Most auth solutions don’t do this out of the box. You’ll need a custom key management layer.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;4-rate-limiting&quot;&gt;4. Rate limiting&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Per-key rate limiting with at least:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;RPM&lt;/strong&gt; (requests per minute)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;TPM&lt;/strong&gt; (tokens per minute)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Budget caps&lt;/strong&gt; (dollar amount per time window)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This needs to be enforced at the proxy layer, before the request hits the inference backend. Otherwise a single user can exhaust your GPU allocation.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Options:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Redis-based counters (most common)&lt;/li&gt;
&lt;li&gt;Token bucket algorithms&lt;/li&gt;
&lt;li&gt;Proxy-level enforcement (some gateways include this)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If you’re using per-key plans, each key needs its own set of limits. Not one global limit — individual limits per key.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;5-usage-tracking-and-billing&quot;&gt;5. Usage tracking and billing&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;You need to know:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;How many tokens each key consumed (input + output)&lt;/li&gt;
&lt;li&gt;What model was used&lt;/li&gt;
&lt;li&gt;Cost per request&lt;/li&gt;
&lt;li&gt;Aggregate usage per user, per day, per billing period&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;For subscription billing:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Stripe for card payments&lt;/li&gt;
&lt;li&gt;Budget windows (e.g., $X per 5-hour period)&lt;/li&gt;
&lt;li&gt;Automatic key revocation when subscription expires&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;For pay-as-you-go:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Credit balance per user&lt;/li&gt;
&lt;li&gt;Deduct per request based on token count × model price&lt;/li&gt;
&lt;li&gt;Top-up flow (Stripe, crypto, etc.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;For crypto payments:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;USDC on a supported chain&lt;/li&gt;
&lt;li&gt;On-chain transaction verification&lt;/li&gt;
&lt;li&gt;Wallet connector in the dashboard (wagmi, viem, etc.)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is a significant amount of code. Usage tracking alone requires intercepting every response to count tokens, calculating cost based on the model’s pricing, and storing it per key.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;6-dashboard&quot;&gt;6. Dashboard&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Your users need a web UI to:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Create and manage API keys&lt;/li&gt;
&lt;li&gt;View usage per key (tokens, requests, cost)&lt;/li&gt;
&lt;li&gt;Subscribe to plans or top up credits&lt;/li&gt;
&lt;li&gt;See available models and pricing&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Tech stack typically:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;React/Next.js/Vue frontend&lt;/li&gt;
&lt;li&gt;REST API backend&lt;/li&gt;
&lt;li&gt;Real-time usage updates&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;For platforms (your users creating keys for their users), you also need a &lt;strong&gt;management API&lt;/strong&gt; — programmatic key creation, plan assignment, usage queries.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;7-model-catalog-management&quot;&gt;7. Model catalog management&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Models change. New ones come out weekly. You need:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A catalog of which models you serve&lt;/li&gt;
&lt;li&gt;Pricing per model (input/output cost per token)&lt;/li&gt;
&lt;li&gt;Sync mechanism to update prices when providers change them&lt;/li&gt;
&lt;li&gt;Display names, categories, tags for the dashboard&lt;/li&gt;
&lt;li&gt;Cache pricing metadata (some models support prompt caching discounts)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is an ongoing operational burden, not a one-time setup.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;8-documentation&quot;&gt;8. Documentation&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Your users need:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;API reference (endpoints, request/response formats)&lt;/li&gt;
&lt;li&gt;SDK examples (Python, Node.js, at minimum)&lt;/li&gt;
&lt;li&gt;Authentication guide&lt;/li&gt;
&lt;li&gt;Billing/usage documentation&lt;/li&gt;
&lt;li&gt;Quick start guide&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is easily 20-30 pages of documentation that needs to stay current.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;9-monitoring-and-reliability&quot;&gt;9. Monitoring and reliability&lt;/h2&gt;&lt;/div&gt;
&lt;ul&gt;
&lt;li&gt;Health checks on the inference backend&lt;/li&gt;
&lt;li&gt;Status page for users&lt;/li&gt;
&lt;li&gt;Alerting when latency spikes or errors increase&lt;/li&gt;
&lt;li&gt;Logging (but not logging prompt content — privacy)&lt;/li&gt;
&lt;li&gt;Graceful degradation when a model or provider is down&lt;/li&gt;
&lt;/ul&gt;
&lt;div&gt;&lt;h2 id=&quot;10-compliance-and-privacy&quot;&gt;10. Compliance and privacy&lt;/h2&gt;&lt;/div&gt;
&lt;ul&gt;
&lt;li&gt;Privacy policy&lt;/li&gt;
&lt;li&gt;Data handling documentation&lt;/li&gt;
&lt;li&gt;GDPR compliance if you serve EU users&lt;/li&gt;
&lt;li&gt;Decision: do you store prompts? (You shouldn’t)&lt;/li&gt;
&lt;li&gt;SOC 2 / ISO 27001 if targeting enterprise&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;div&gt;&lt;h2 id=&quot;the-full-stack&quot;&gt;The full stack&lt;/h2&gt;&lt;/div&gt;

















































&lt;table&gt;&lt;thead&gt;&lt;tr&gt;&lt;th&gt;Component&lt;/th&gt;&lt;th&gt;Ongoing maintenance&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td&gt;Inference backend&lt;/td&gt;&lt;td&gt;High — scaling, failover, model updates&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;API proxy&lt;/td&gt;&lt;td&gt;Medium — format changes, new providers&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Auth + key management&lt;/td&gt;&lt;td&gt;Low&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Per-key rate limiting&lt;/td&gt;&lt;td&gt;Low&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Usage tracking + billing&lt;/td&gt;&lt;td&gt;Medium — edge cases, reconciliation&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Dashboard&lt;/td&gt;&lt;td&gt;Medium — new features, UX&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Model catalog&lt;/td&gt;&lt;td&gt;High — weekly model updates&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Documentation&lt;/td&gt;&lt;td&gt;Medium — keep current&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Monitoring&lt;/td&gt;&lt;td&gt;Low&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td&gt;Privacy/compliance&lt;/td&gt;&lt;td&gt;Low&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;
&lt;div&gt;&lt;h2 id=&quot;what-breaks-in-production&quot;&gt;What breaks in production&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;Building is the easy part. The hard part is what breaks with real users:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A provider changes their API format without warning. Your proxy returns 500s for 2 hours until you notice.&lt;/li&gt;
&lt;li&gt;A model gets deprecated. Your users’ hardcoded model IDs stop working overnight.&lt;/li&gt;
&lt;li&gt;Token counting has an off-by-one bug. You’ve been undercharging for 3 weeks. Your margin is gone.&lt;/li&gt;
&lt;li&gt;A user finds a way to exceed rate limits through concurrent requests. Your inference bill spikes 10x in one afternoon.&lt;/li&gt;
&lt;li&gt;Stripe webhook fails silently. A user’s subscription expired but their API key still works. Free inference for a month.&lt;/li&gt;
&lt;li&gt;You push a billing update and break the usage tracking. Three days of missing data. Users open tickets.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Each of these has happened to us. We fixed them. The question is whether you want to fix them yourself, with your users waiting, or use a platform that already has.&lt;/p&gt;
&lt;div&gt;&lt;h2 id=&quot;or&quot;&gt;Or&lt;/h2&gt;&lt;/div&gt;
&lt;p&gt;You use an inference platform that already has all of this, create API keys for your users, and ship your product this week.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;We built all of the above so you don’t have to. &lt;a href=&quot;https://cheapestinference.com/platforms&quot;&gt;See how per-key plans work&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;</content:encoded></item></channel></rss>