DeepInfra raises $107M Series B to scale the inference cloud — read the announcement

While everyone knows Llama 3 and Qwen, a quieter revolution has been happening in NVIDIA’s labs. They have been taking standard Llama models and “supercharging” them using advanced alignment techniques and pruning methods.
The result is Nemotron—a family of models that frequently tops the “Helpfulness” leaderboards (like Arena Hard), often beating GPT-4o while being significantly more efficient to run.
NVIDIA’s strategy is unique: they don’t just train models; they optimize them for hardware. This means you get models like the Nemotron-Super-49B, which delivers 70B-level intelligence at a fraction of the cost and memory footprint.
This guide breaks down the pricing for the Nemotron family on DeepInfra and helps you decide which one fits your budget.
If you are new to LLM APIs, the pricing can look confusing. You aren’t paid by the request or by the minute; you are charged by the “Token”.
Here is the simple breakdown of how to calculate your costs:
DeepInfra offers the full range of NVIDIA’s Nemotron models. Because these models are optimized for NVIDIA hardware (which DeepInfra runs on), the pricing is often very aggressive, especially for the “Super” and “Nano” variants.
You can view the full list and test them here: DeepInfra Nemotron Models.
| Model Name | Context Window | Input Price (per 1M) | Output Price (per 1M) |
| Llama-3.3-Nemotron-Super-49B-v1.5 | 128K | $0.10 | $0.40 |
| Llama-3.1-Nemotron-70B-Instruct | 128K | $1.20 | $1.20 |
| NVIDIA-Nemotron-Nano-12B-v2-VL | 128K | $0.20 | $0.60 |
| NVIDIA-Nemotron-Nano-9B-v2 | 128K | $0.04 | $0.16 |
Note: Prices are per 1 million tokens. A 128K context window allows these models to process entire books or long codebases in a single prompt.
The most interesting model on this list is undoubtedly the Llama-3.3-Nemotron-Super-49B.
Typically, to get “70B level” performance, you have to pay for a 70B parameter model. NVIDIA used a technique called Neural Architecture Search (NAS) to take the Llama 3.3 70B model and intelligently prune (remove) the parts of the brain that weren’t contributing much to intelligence.
If you are building a RAG application or a chatbot, the Super-49B is likely the “sweet spot” for 2025.
You might notice the Llama-3.1-Nemotron-70B-Instruct is significantly more expensive at $1.20/$1.20. Why?
This model wasn’t pruned for speed; it was optimized for quality. NVIDIA trained this using a special “HelpSteer2” dataset and advanced Reinforcement Learning from Human Feedback (RLHF).
While the base Llama 3.1 is smart, the Nemotron version is “better behaved.” It is less likely to refuse requests, gives more structured answers, and scores higher on “human preference” benchmarks. You pay a premium for this polish. It is best used for client-facing outputs where tone and strict instruction following are critical.
Let’s see how much you would actually save by choosing the right Nemotron model.
Estimated Cost:
(If you used the standard Nemotron 70B for this, the bill would be roughly $66.00. The “Super” model saves you nearly 90%.)
Estimated Cost:
At $0.20 per million input tokens, this is one of the most affordable Vision-Language models on the market. Competitors like GPT-4o charge upwards of $2.50 for similar multimodal inputs.
The Nemotron family offers a unique value proposition: NVIDIA-grade optimization on top of Meta’s open weights.
By selecting the specific Nemotron variant optimized for your workload, you can achieve better-than-GPT-4o results while keeping your infrastructure costs incredibly low.
Reliable JSON-Only Responses with DeepInfra LLMs<p>When large language models are used inside real applications, their role changes fundamentally. Instead of chatting with users, they become infrastructure components: extracting information, transforming text, driving workflows, or powering APIs. In these scenarios, natural language is no longer the desired output. What applications need is structured data — and very often, that structure is […]</p>
How to use CivitAI LoRAs: 5-Minute AI Guide to Stunning Double Exposure ArtLearn how to create mesmerizing double exposure art in minutes using AI. This guide shows you how to set up a LoRA model from CivitAI and create stunning artistic compositions that blend multiple images into dreamlike masterpieces.
DeepInfra is now a supported Hugging Face Inference ProviderDeepInfra is officially live as an Inference Provider on the Hugging Face Hub. You can now call DeepInfra-hosted models directly from Hugging Face model pages, through our OpenAI-compatible router (use it with any OpenAI SDK), or via the Hugging Face SDKs in Python and JavaScript.© 2026 DeepInfra. All rights reserved.