Today, we are releasing the latest version of our flagship model: GLM-4.6. Compared with GLM-4.5, this generation brings several key improvements:
We evaluated GLM-4.6 across eight public benchmarks covering agents, reasoning, and coding. Results show clear gains over GLM-4.5, with GLM-4.6 also holding competitive advantages over leading domestic and international models such as DeepSeek-V3.2-Exp and Claude Sonnet 4, but still lags behind Claude Sonnet 4.5 in coding ability.
Real-world experience matters more than leaderboards. We extended CC-Bench from GLM-4.5 with more challenging tasks, where human evaluators worked with models inside isolated Docker containers and completed multi-turn real-world tasks across front-end development, tool building, data analysis, testing, and algorithm. GLM-4.6 improves over GLM-4.5 and reaches near parity with Claude Sonnet 4 (48.6% win rate), while clearly outperforming other open-source baselines. From a token-efficiency perspective, GLM-4.6 finishes tasks with about 15% fewer tokens than GLM-4.5, showing improvements in both capability and efficiency. All evaluation details and trajectory data have been made publicly available for further community research: https://huggingface.co/datasets/zai-org/CC-Bench-trajectories
The Z.ai API platform offers both GLM-4.6 models. For comprehensive API documentation and integration guidelines, please refer to https://docs.z.ai/guides/llm/glm-4.6. Alternatively, developers are welcome to access both models through OpenRouter.
GLM-4.6 is now available to use within coding agents (Claude Code, Kilo Code, Roo Code, Cline and more).
For GLM Coding Plan subscribers: You'll be automatically upgraded to GLM-4.6. If you've previously customized the app configs (like ~/.claude/settings.json
in Claude Code), simply update the model name to "glm-4.6" to complete the upgrade.
For New users: The GLM Coding Plan offers Claude-level performance at a fraction of the cost — just 1/7th the price with 3x the usage quota. Start building today: https://z.ai/subscribe.
GLM-4.6 is accessible through Z.ai by selecting the GLM-4.6 model option.
Model weights of GLM-4.6 will soon be available at HuggingFace and ModelScope. For local deployment, GLM-4.6 supports inference frameworks including vLLM and SGLang. Comprehensive deployment instructions are available in the official GitHub repository.