Requesty
Z AI

GLM-5.1

Compared with GLM-5, GLM-5.1 delivers significant improvements in coding, agentic tool usage, reasoning, role-play, and general chat quality. Besides, GLM-5.1 has outstanding capabilities in long-horizon agentic tasks like CUDA kernel optimization.

πŸ‘Vision🧠ReasoningπŸ”§Tool calling⚑Caching

Pricing per 1M tokens

Input
$1.40
Output
$4.40
Cache write
$4.40
Cache read
$0.26

Specifications

Context window200K tokens
Max output128K tokens
API typechat
AddedApr 7, 2026
Model IDzai/GLM-5.1

Privacy & data

Data retentionNo
Used for trainingNo
Provider locationπŸ‡ΈπŸ‡¬ Singapore