Try GLM 5 in the Workbench
Run this model interactively, tune parameters, and compare outputs.
zai-org-glm-5
zai-org/GLM-5 is a frontier-level large language model developed by Z AI with exceptional scores across benchmarks rivaling the latest GPT and Claude models. It scales to 744 billion total parameters with 40 billion active parameters per token, featuring DeepSeek Sparse Attention (DSA) for efficient long-context inference. GLM-5 excels at coding, reasoning, and long-horizon agentic tasks, ranking #1 among open-source models on Vending Bench 2 (long-term operational planning) and achieving best-in-class performance on reasoning benchmarks and coding tasks like SWE-bench.
| Metric | Value |
|---|---|
| Parameter Count | 744 billion |
| Mixture of Experts | Yes |
| Active Parameter Count | 40 billion |
| Context Length | 200,000 tokens |
| Multilingual | Yes |
| Quantized* | Yes |
| Precision* | NVFP4 |
Example request
- Minimal
- Basic parameters
- All parameters
Fetch model details
The models endpoint returns the full model object, including itsjson_request_schema.