Z.AI: GLM 4.6
Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks. Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Code、Cline、Roo Code and Kilo Code, including improvements in generating visually polished front-end pages. Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability. More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks. Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios.
Model Information
Pricing Information
Supported Parameters
Common Use Cases
Text Generation
- • Content writing and editing
- • Code generation and debugging
- • Creative writing and storytelling
- • Translation and summarization
General Applications
- • Chatbots and virtual assistants
- • Educational content creation
- • Research and analysis
- • Automation and workflow
Frequently Asked Questions
What is the context length of this model?
This model has a context length of 203K tokens, which means it can process and remember up to 203K tokens of text in a single conversation or request.
How much does it cost to use this model?
Prompt tokens cost $0.50M/1M tokens and completion tokens cost $1.75M/1M tokens.
What modalities does this model support?
This model supports text->text modality, accepting textas input and producing text as output.
When was this model created?
This model was created on September 30, 2025.