Skip to content

REST API

OpenAI 兼容的 REST API。

端点

POST /v1/chat/completions

创建聊天补全。兼容 OpenAI SDK。

请求:

json
{
  "model": "qwen/qwen-turbo",
  "messages": [
    { "role": "system", "content": "你是一个有帮助的助手。" },
    { "role": "user", "content": "你好!" }
  ],
  "stream": false,
  "temperature": 0.7,
  "max_tokens": 2048
}

响应:

json
{
  "id": "chatcmpl-xxx",
  "object": "chat.completion",
  "created": 1234567890,
  "model": "qwen/qwen-turbo",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "你好!有什么可以帮你的?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

流式输出(SSE):

设置 "stream": true 接收 Server-Sent Events:

bash
curl -N http://localhost:18888/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{"model": "default", "messages": [{"role": "user", "content": "你好"}], "stream": true}'

GET /v1/models

列出可用模型。

GET /health

健康检查端点。

GET /status

网关状态,包含运行时间、连接的通道和活跃会话。

POST /hooks/:name

Webhook 入口,用于自定义集成。

使用 OpenAI SDK

python
from openai import OpenAI

client = OpenAI(
    base_url="http://localhost:18888/v1",
    api_key="not-needed",
)

response = client.chat.completions.create(
    model="default",
    messages=[{"role": "user", "content": "你好!"}],
)
print(response.choices[0].message.content)
typescript
import OpenAI from 'openai';

const client = new OpenAI({
  baseURL: 'http://localhost:18888/v1',
  apiKey: 'not-needed',
});

const response = await client.chat.completions.create({
  model: 'default',
  messages: [{ role: 'user', content: '你好!' }],
});

Released under MIT / Apache-2.0 License