Ollama Pool

AI Model Gateway & Load Balancer

Welcome

Set up your admin account to get started

Instance Status

NameHostStatusModels

Model Speed Ranking

ModelBest SpeedAvg SpeedAvailable

Ollama Management

NameHostStatusModelsActions

Models

ModelDescTypeSizeAvailableBest SpeedAction

Official API Key

NameAPI KeyStatusActions

By Model

ModelRequestsTokensAvg Latency

By API Key

KeyRequestsTokensAvg Latency

API Keys

NameKeyStatusCreatedActions

API Documentation

Base URL:

OpenAI Format
Anthropic Format

POST /v1/chat/completions

OpenAI 兼容聊天补全接口,支持流式输出。

请求参数

参数类型必填默认值说明
modelstring-模型名称,如 llama3.2、kimi-k2.6:cloud
messagesarray-对话历史,每项含 role 和 content
temperaturefloat0.5温度 (0-2),控制随机性
max_tokensint无限制最大生成 token 数
streamboolfalse是否流式输出
top_pfloat-Top-p 采样 (0-1)
frequency_penaltyfloat-频率惩罚 (-2 到 2)
presence_penaltyfloat-存在惩罚 (-2 到 2)
repetition_penaltyfloat1.2重复惩罚 (Ollama 特有)
stoparray-停止序列列表

请求示例

curl ${location.origin}/v1/chat/completions \
  -H "Authorization: Bearer sk-your-api-key" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "llama3.2",
    "messages": [{"role": "user", "content": "Hello!"}],
    "temperature": 0.7,
    "max_tokens": 1024,
    "stream": false
  }'

响应格式

{
  "id": "chatcmpl-xxx",
  "object": "chat.completion",
  "model": "llama3.2",
  "choices": [{
    "index": 0,
    "message": {"role": "assistant", "content": "..."},
    "finish_reason": "stop"
  }],
  "usage": {"prompt_tokens": 5, "completion_tokens": 10, "total_tokens": 15}
}

流式响应 (stream: true)

data: {"choices":[{"delta":{"role":"assistant"}}]}
data: {"choices":[{"delta":{"content":"Hello"}}]}
data: {"choices":[{"delta":{"content":"!"}}]}
data: {"choices":[{"delta":{},"finish_reason":"stop"}]}
data: [DONE]

扩展:支持推理的模型会发送 reasoning_content 字段:

data: {"choices":[{"delta":{"reasoning_content":"让我思考..."}}]}

GET /v1/models

curl ${location.origin}/v1/models \
  -H "Authorization: Bearer sk-your-api-key"
{
  "object": "list",
  "data": [
    {"id": "llama3.2", "object": "model", "owned_by": "ollama"},
    {"id": "kimi-k2.6:cloud", "object": "model", "owned_by": "cloud"}
  ]
}

Python SDK 示例

from openai import OpenAI

client = OpenAI(
    api_key="sk-your-api-key",
    base_url="${location.origin}/v1",
)

# 非流式
resp = client.chat.completions.create(
    model="llama3.2",
    messages=[{"role": "user", "content": "Hello!"}],
    temperature=0.7,
    max_tokens=1024,
)
print(resp.choices[0].message.content)

# 流式
stream = client.chat.completions.create(
    model="llama3.2",
    messages=[{"role": "user", "content": "Hello!"}],
    stream=True,
)
for chunk in stream:
    if chunk.choices[0].delta.content:
        print(chunk.choices[0].delta.content, end="")
    if chunk.choices[0].delta.reasoning_content:
        print(chunk.choices[0].delta.reasoning_content, end="")

Database Backup

Export SQLite database file for backup, or restore from a previous backup.

Invite Codes

CodeUsesMaxExpiresStatusActions

Users

EmailRoleStatusJoinedLast LoginActions