DeepInfra raises $107M Series B to scale the inference cloud — read the announcement

GLM-5 is the latest open-weights reasoning model released by Z AI (Zhipu AI) in February 2026, characterized by high “thinking token” usage. It is a Mixture of Experts (MoE) model with 744B total parameters and 40B active parameters, scaling up from GLM-4.5’s 355B parameters. The model was pre-trained on 28.5T tokens and features a 200K+ context window.
GLM-5 is purpose-built for complex systems engineering and long-horizon agentic tasks. It integrates DeepSeek Sparse Attention (DSA), substantially reducing deployment costs while preserving long-context capacity. The model was trained using a novel asynchronous RL infrastructure called “Slime” that improves training throughput and efficiency.
GLM-5 achieves best-in-class performance among all open-source models on reasoning, coding, and agentic tasks, closing the gap with frontier models like Claude Opus 4.5. On SWE-bench Verified and Terminal-Bench 2.0, GLM-5 records leading open-model scores of 77.8 and 56.2, respectively.
GLM-5 is now available across multiple inference providers, but they’re not created equal. This analysis breaks down which one delivers the best performance, lowest cost, and fastest response times for your use case.
| Provider | Best For | Blended ($/1M) | Input ($/1M) | Output ($/1M) | Speed (t/s) | Latency (TTFT) | JSON | Func Calling | Why Notable |
|---|---|---|---|---|---|---|---|---|---|
| DeepInfra FP8 (deepinfra.com) | Lowest cost / best cost baseline | $1.24 | $0.80 | $2.56 | 57 | 1.22s | Yes | Yes | Cheapest blended + input/output; competitive latency for interactive apps at budget. |
| Fireworks | Max throughput + lowest latency | $1.55 | — | — | 212.2 | 0.74s | Yes | Yes | Best raw performance (fastest output + lowest TTFT) but higher blended price. |
| Baseten | Strong balance (fast + low cost) | $1.50 | $0.95 | $3.15 | 183.1 | 0.83s | Yes | Yes | Near-top performance with 2nd-lowest blended price. |
| Eigen AI | High speed alternative | $1.55 | — | — | 204.5 | 1.53s | Yes | Yes | Very high output speed; blended price higher than DeepInfra. |
| Novita FP8 | Price runner-up | $1.55 | $1.00 | $3.20 | 50 | 1.45s | Yes | Yes | Among lowest blended prices, but still above DeepInfra on all metrics. |
Based on benchmarks of all the aforementioned providers, DeepInfra is the recommended API for production-scale GLM-5 deployment. It offers the lowest blended price on the market ($1.24 per 1M tokens), which is critical for reasoning models that generate high volumes of internal chain-of-thought tokens. For use cases requiring sub-second latency, Fireworks is the fastest provider tested.
| Use Case | Recommended Provider | Why? |
|---|---|---|
| Chatbots / Real-time | Fireworks | Lowest Latency (0.74s) |
| Batch Processing / RAG | DeepInfra | Lowest Cost ($1.24/1M) |
| Agentic Workflows | Baseten | Balanced Speed & Cost |
While other providers chase raw burst speed, DeepInfra (FP8) secures the top recommendation by dominating the most critical metric for scaling reasoning models: Cost Efficiency.
Reasoning models like GLM-5 generate a high volume of “thinking” tokens before producing a final answer. This drastically inflates output token usage compared to standard LLMs. DeepInfra offers the lowest blended price and, crucially, the most competitive output pricing, making it the optimal choice for high-volume production environments where margins matter.
DeepInfra provides a competitive latency profile (tying for 4th place) while undercutting the market standard price ($1.55) by 20%. For applications requiring heavy reasoning chains, the savings on output tokens make DeepInfra the superior architectural choice.
For applications requiring massive batch processing or long-form content generation where the user is waiting for the stream to finish, Output Tokens Per Second (t/s) is the governing metric.
Technical Note: The gap between the top tier (Fireworks/Eigen) and the mid-tier (Google at 75.7 t/s) is substantial. If your application relies on rapid text generation (e.g., code autocompletion), the premium for Fireworks or Eigen is justified.
TTFT is the critical metric for conversational AI and chatbots. It measures the time between the request and the first visible character.
DeepInfra and Together.ai (FP4) sit in the second tier at 1.22s. While slower than Fireworks, 1.2 seconds is generally acceptable for most asynchronous reasoning tasks where the user expects a brief thinking pause.
Winner: DeepInfra (FP8)
While speed is important, the cost structure is the deciding factor for reasoning models. GLM-5 generates significant “thinking” tokens, inflating output costs.
DeepInfra undercuts the market average by 20%. For high-volume reasoning chains, this price difference is the primary differentiator. The market has largely coalesced around a standard blended price of $1.55 per 1 million tokens. Deviations from this norm highlight the value leaders:
Technical integration is just as important as raw speed.
| Provider | Blended Price (/1M) | Latency (TTFT) | Output Speed (t/s) | Recommendation |
|---|---|---|---|---|
| DeepInfra FP8 | $1.24 | 1.22s | ~57 | Best Overall (Value) |
| Fireworks | $1.55 | 0.74s | 212.2 | Best Performance |
| Baseten | $1.50 | 0.83s | 183.1 | Best Balanced |
| Eigen AI | $1.55 | 1.53s | 204.5 | High Throughput |
| FriendliAI | $1.55 | 0.94s | 73.9 | Low Latency |
| $1.55 | 1.33s | 75.7 | Standard |
For developers integrating GLM-5 (Reasoning) into their stack, the optimal choice depends on the specific bottleneck of the application.
If you are building a consumer-facing chatbot where every millisecond counts, Fireworks is the technically superior option. However, for the vast majority of enterprise use cases — where reasoning models are used for complex data processing, RAG pipelines, or agentic workflows — DeepInfra is the recommended provider. It offers a robust feature set (JSON mode and Function calling), acceptable latency, and a pricing structure that provides a significant long-term competitive advantage.
Inference Economics: True AI Costs at Scale<p>Most teams discover their inference economics the same way: a production bill arrives that looks nothing like the number they expected. The per-token price seemed small enough during testing. Then real traffic showed up, agents started chaining calls, RAG pipelines bloated the context window, and suddenly the math looked completely different. Token prices have fallen […]</p>
Qwen3.5 9B API Benchmarks: Latency, Throughput & Cost<p>About Qwen3.5 9B Qwen3.5 9B is the flagship of Alibaba’s Qwen3.5 Small Model Series, released on March 2, 2026. It is a dense multimodal model combining Gated Delta Networks (a form of linear attention) with a sparse Mixture-of-Experts system, enabling higher throughput and lower latency during inference compared to traditional dense architectures. The architecture utilizes […]</p>
Fork of Text Generation Inference.The text generation inference open source project by huggingface looked like a promising
framework for serving large language models (LLM). However, huggingface announced that they
will change the license of code with version v1.0.0. While the previous license Apache 2.0
was permissive, the new on...© 2026 DeepInfra. All rights reserved.