Chinese AI developer Zhipu AI, with support from ByteDance, has released the technical report and open-source code for GLM-4.5, a Mixture-of-Experts large language model aimed at reasoning, coding and autonomous-agent tasks. The system contains 355 billion parameters, with 32 billion activated during inference, and was trained on 23 trillion tokens using a multi-stage regimen that combines expert iteration, self-distillation and reinforcement learning. According to the report, GLM-4.5 achieves 70.1 % on the TAU-Bench reasoning test, 91.0 % on the AIME 24 math benchmark and 64.2 % on the SWE-bench Verified software-engineering suite. The scores place the model third overall across 12 widely used evaluation sets and second on dedicated agentic benchmarks, while using fewer active parameters than many competing systems. Zhipu AI also released a smaller 106-billion-parameter version, GLM-4.5-Air, as well as GLM-4.5V, a vision-language variant that reports state-of-the-art results on 42 image, video and document-understanding tasks. The models and accompanying code are available on Hugging Face and GitHub under open-source licenses to encourage further research and commercial adoption.
ZhipuAI just released GLM-4.5V on Hugging Face A new vision-language model with advanced reasoning, achieving SOTA performance on 42 benchmarks for image, video, and document understanding. https://t.co/y8IoIZ4wYv
technical report GLM-4.5, the tl;dr -ARC focus: The model is designed as an “Agentic, Reasoning, and Coding” (ARC) base model and aims to achieve consistent high performance in these three critical areas. -Efficient MoE architecture: It uses a “Mixture-of-Experts” (MoE) https://t.co/3do9MLvpAK https://t.co/0QGSYhNB8a
GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models "Through multi-stage training on 23T tokens and comprehensive post-training with expert model iteration and reinforcement learning, GLM-4.5 achieves strong performance across agentic, reasoning, and coding (ARC) https://t.co/pTCqZpFzb8