Zhinao API routes requests to the best-fit provider and automatically fails over to the one with highest availability.
TTFT
No data
Throughput
No data
Uptime
No data
Provider Model
volcengine/deepseek-v3-250324
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
65,536
Max Output
8,096
Input Price
¥2 / 1M tokens
Output Price
¥8 / 1M tokens
TTFT
0.06s
Throughput
21.01tps
Uptime
100.00%
Provider Model
qiniu/deepseek-v3
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
65,536
Max Output
8,096
Input Price
¥1.3 / 1M tokens
Output Price
¥5.2 / 1M tokens
TTFT
No data
Throughput
20.83tps
Uptime
100.00%
Provider Model
baidu/deepseek-v3
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
65,536
Max Output
8,096
Input Price
¥1 / 1M tokens
Output Price
¥4 / 1M tokens
TTFT
4.85s
Throughput
12.02tps
Uptime
100.00%
Provider Model
siliconflow/deepseek-v3
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
65,536
Max Output
8,096
Input Price
¥2 / 1M tokens
Output Price
¥8 / 1M tokens
TTFT
0.35s
Throughput
2.41tps
Uptime
100.00%
Provider Model
tencent/deepseek-v3
Supported Parameters
Recent Uptime
Reasoning
-
Supported Response Formats
Request Log Collection
-
Distillable
-
Total Context
65,536
Max Output
8,096
Input Price
¥2 / 1M tokens
Output Price
¥8 / 1M tokens
Compare different providers across Zhinao API
18.64 tok/s
1.77 s
Uptime for deepseek-chat-v3 across all providers
Zhinao API normalizes requests and responses across providers for you
import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.360.cn/v1",
apiKey: process.env.ZHINAO_API_KEY,
});
const response = await client.chat.completions.create({
model: "deepseek-chat-v3",
messages: [
{ role: "user", content: "Hello, how are you?" }
],
temperature: 0.7,
max_tokens: 1000,
});
console.log(response.choices[0].message.content);