Publish time
2025/7/28Model Series
GLMInput type
Output type
Context Window
128,000Max Output Length
98,304Input Price
¥2 / 1M tokensOutput Price
¥8 / 1M tokensGLM-4.5 是一款专为智能体应用打造的基础模型,使用了混合专家(Mixture-of-Experts)架构。在工具调用、网页浏览、软件工程、前端编程领域进行了深度优化,支持无缝接入 Claude
Zhinao API routes requests to the best-fit provider and automatically fails over to the one with highest availability.
TTFT
5.99s
Throughput
34.90tps
Uptime
97.00%
Provider Model
bigmodel/glm-4.5
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
128,000
Max Output
98,304
Input Price
¥2 / 1M tokens
Output Price
¥8 / 1M tokens
Compare different providers across Zhinao API
19.67 tok/s
0.91 s
Uptime for z-ai/glm-4.5 across all providers
Zhinao API normalizes requests and responses across providers for you
import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.360.cn/v1",
apiKey: process.env.ZHINAO_API_KEY,
});
const response = await client.chat.completions.create({
model: "z-ai/glm-4.5",
messages: [
{ role: "user", content: "Hello, how are you?" }
],
temperature: 0.7,
max_tokens: 1000,
});
console.log(response.choices[0].message.content);