Z1: Efficient Test-time Scaling with Code

Train Large Language Model to Reason with Shifted Thinking

[๐Ÿ“œ Paper] โ€ข [๐Ÿค— HF Models] โ€ข [๐Ÿฑ GitHub]

Model Details

To begin with the shifted thinking mode, please refer to https://github.com/efficientscaling/Z1.

Evaluation


Citation

@misc{yu2025efficientscaling,
      title={Z1: Efficient Test-time Scaling with Code}, 
      author={Zhaojian Yu and Yinghao Wu and Yilun Zhao and Arman Cohan and Xiao-Ping Zhang},
      year={2025},
      eprint={2504.00810},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2504.00810}, 
}
Downloads last month
38
Safetensors
Model size
7.61B params
Tensor type
FP16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for efficientscaling/Z1-7B

Base model

Qwen/Qwen2.5-7B
Finetuned
(75)
this model
Quantizations
2 models