Released Apr 7, 2026202,752 context$1.26/M input tokens$3.96/M output tokens
GLM-5.1 delivers a major leap in coding capability, with particularly significant gains in handling long-horizon tasks. Unlike previous models built around minute-level interactions, GLM-5.1 can work independently and continuously on a single task for more than 8 hours, autonomously planning, executing, and improving itself throughout the process, ultimately delivering complete, engineering-grade results.
Recent activity on GLM 5.1
Total usage per day on OpenRouter
Prompt
47.1B
Reasoning
367M
Completion
340M
Prompt tokens measure input size. Reasoning tokens show internal thinking before a response. Completion tokens reflect total output length.