GLM-5.1 - state-of-the-art agentic engineering, now available on DeepInfra!

Recently an interesting new model got released. It is called Lzlv, and it is basically a merge of few existing models. This model is using the Vicuna prompt format, so keep this in mind if you are using our raw API. If you are using the OpenAI API we will take care of the prompt formatting for you.
The Lzlv model also got lots of attention on Reddit, and it was tagged as top choice for the 70b category here: https://www.reddit.com/r/LocalLLaMA/comments/17fhp9k/huge_llm_comparisontest_39_models_tested_7b70b/ It seems to have the right balance between creativity and coherence and is a good choice for roleplay use-cases. You can quickly try it out using our web chat ui. If you like the results you can easily integrate it into your application using our APIs.
As always you can easily access the model using our LLM APIs. For example using our OpenAI compatible API:
import openai
# Point OpenAI client to our endpoint, get api key: https://deepinfra.com/dash/api_keys
openai.api_key = "<YOUR DEEPINFRA TOKEN>"
openai.api_base = "https://api.deepinfra.com/v1/openai"
# Your chosen model here
MODEL_DI = "lizpreciatior/lzlv_70b_fp16_hf"
chat_completion = openai.ChatCompletion.create(
model=MODEL_DI,
messages=[{"role": "user", "content": "Hello there"}],
stream=True,
max_tokens=100,
)
# print the chat completion
for event in chat_completion:
print(event.choices)
You can also use our lower level HTTP API directly. This will allow you to build your own prompt, and integrate it easily into any application.
curl -X POST \
-d '{"input": "Tell me a joke."}' \
-H "Authorization: bearer <YOUR DEEPINFRA API TOKEN>" \
-H 'Content-Type: application/json' \
'https://api.deepinfra.com/v1/inference/lizpreciatior/lzlv_70b_fp16_hf'
We have both streaming and non streaming APIs available and you can find more details here: https://deepinfra.com/lizpreciatior/lzlv_70b_fp16_hf/api
This is a 70b model, and it is quite expensive to run. Typically, you need at least couple A100 80GB GPUs to run it, but to get the best performance you need at least 4 high end GPUs. Getting this hardware and setting everything up is quite expensive and time-consuming. Our APIs however are the most cost-effective way to run this model. For 70b models we charge 0.7 USD per 1M input tokens, and 0.9 USD per 1M output tokens. The biggest saving is that you only pay for usage and not the time.
If you need any help, just reach out to us on our Discord server.
How to use CivitAI LoRAs: 5-Minute AI Guide to Stunning Double Exposure ArtLearn how to create mesmerizing double exposure art in minutes using AI. This guide shows you how to set up a LoRA model from CivitAI and create stunning artistic compositions that blend multiple images into dreamlike masterpieces.
Qwen3.5 397B A17B API Benchmarks: Latency, Throughput & Cost<p>About Qwen3.5 397B A17B Qwen3.5 397B A17B is Alibaba Cloud’s largest and most capable multimodal foundation model, released in February 2026. It features a hybrid Mixture-of-Experts (MoE) architecture with 397 billion total parameters and 17 billion active parameters per inference pass, utilizing 512 experts with a routing mechanism selecting a subset per token. This sparse […]</p>
Qwen API Pricing Guide 2026: Max Performance on a Budget<p>If you have been following the AI leaderboards lately, you have likely noticed a new name constantly trading blows with GPT-4o and Claude 3.5 Sonnet: Qwen. Developed by Alibaba Cloud, the Qwen model family (specifically Qwen 2.5 and Qwen 3) has exploded in popularity for one simple reason: unbeatable price-to-performance. In 2025, Qwen is widely […]</p>
© 2026 Deep Infra. All rights reserved.