GLM-4.7-Flash is Z.ai’s 30B MoE model built for low-latency reasoning and tool calling—plus benchmarks like AIME 2025, GPQA, and SWE-bench.
GLM-4.7-Flash is Z.ai’s 30B MoE model built for low-latency reasoning and tool calling—plus benchmarks like AIME 2025, GPQA, and SWE-bench.