|
--- |
|
license: mit |
|
language: |
|
- zh |
|
- en |
|
pipeline_tag: text-generation |
|
library_name: transformers |
|
--- |
|
|
|
# GLM-4-32B-0414 |
|
|
|
## Introduction |
|
|
|
Based on our latest technological advancements, we have trained a `GLM-4-0414` series model. During pretraining, we incorporated more code-related and reasoning-related data. In the alignment phase, we optimized the model specifically for agent capabilities. As a result, the model's performance in agent tasks such as tool use, web search, and coding has been significantly improved. |
|
|
|
 |
|
|
|
| Models | IFEval | SWE-Bench | BFCL-v3 (Overall) | BFCL-v3 (MultiTurn) | TAU-Bench (Retail) | TAU-Bench (Airline) | SimpleQA | HotpotQA | |
|
|------------------|---------|------------------|-------------------|---------------------|--------------------|---------------------|----------|----------| |
|
| Qwen2.5-Max | 85.6 | 24.4 | 50.9 | 30.5 | 58.3 | 22.0 | 79.0 | 52.8 | |
|
| GPT-4o-1120 | 81.9 | 38.8 | 69.6 | 41.0 | 62.8 | 46.0 | 82.8 | 63.9 | |
|
| DeepSeek-V3-0324 | 83.4 | 38.8(oh) | 66.2 | 35.8 | 60.7 | 32.4 | 82.6 | 54.6 | |
|
| DeepSeek-R1 | 84.3 | 34(oh)/ 49.2(al) | 57.5 | 12.4 | 33.0 | 37.3 | 83.9 | 63.1 | |
|
| GLM-4-32B-0414 | 86.5 | | 69.6 | 41.5 | 68.7 | 51.2 | 88.1 | 63.8 | |
|
|
|
|
|
## Inference Code |
|
|
|
Make Sure Using `transforemrs>=4.51.3`. |
|
|
|
This model is a base model. If you need a chat model, please use the [GLM-4-32B-Chat-0414](https://huggingface.co/THUDM/GLM-4-32B-Chat-0414) chat model. |