GLM 5 Turbo

Zhipu
zhipu/glm-5-turbo

Reasoning model, 131K context

Context Window

131K

131,072 tokens

Max Output

16K

16,384 tokens

About this model

GLM-5 Turbo β€” fast and cost-effective coding model

This model supports up to 131K tokens of context. It excels at complex reasoning, mathematical problems, and multi-step tasks. It provides strong code generation and debugging capabilities.

Access it through Chuizi.AI with a single ck- API key β€” no separate Zhipu AI account needed.

Highlights

131K context window
16K max output
Advanced reasoning
Strong code generation

Best For

Complex coding tasksAlgorithm designCode reviewTechnical architecture
2026-01-01

Capabilities

ChatReasoningCodetools

Aliases

glm-5-turbo

Pricing (per 1M tokens)

Pricing (per 1M tokens)/ 1M tokens
Input / 1M$0.59
Output / 1M$2.31

Final prices shown

Quick Start

main.py
from openai import OpenAI

client = OpenAI(
    base_url="https://api.chuizi.ai/v1",
    api_key="ck-your-key-here",
)

response = client.chat.completions.create(
    model="zhipu/glm-5-turbo",
    messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)

FAQ

Related Models

GLM 5 Turbo β€” Pricing, Context, Capabilities | Chuizi AI