Skip to main content
POST
/
v1
/
chat
/
completions
curl -X POST https://qingbo.dev/v1/chat/completions \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "gpt-5",
    "messages": [
      {
        "role": "system",
        "content": "你是一个专业的AI助手。"
      },
      {
        "role": "user",
        "content": "解释一下冒泡排序算法。"
      }
    ],
    "temperature": 0.7,
    "max_tokens": 1024
  }'
{
  "id": "chatcmpl-CdA8u0qPqzeN2gu8QLuQfjaP9CHU4",
  "object": "chat.completion",
  "created": 1763449816,
  "model": "gpt-5-2025-08-07",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "冒泡排序(Bubble Sort)是一种简单的比较交换排序。它通过重复遍历序列,比较相邻元素并把较大的"冒泡"到右侧末端,直到序列有序。\n\n核心思想\n- 每一趟从左到右比较相邻元素,若前者大于后者则交换。一次趟结束后,当前未排部分中最大的元素被放到末尾。\n- 重复多趟,直到没有发生交换或已完成 n−1 趟。\n\n流程示例(以 [5, 1, 4, 2] 为例)\n- 第1趟:比较并交换 → [1, 5, 4, 2] → [1, 4, 5, 2] → [1, 4, 2, 5](最大元素5到末尾)\n- 第2趟:比较并交换 → [1, 4, 2, 5] → [1, 2, 4, 5](已整体有序,可提前结束)\n\n性质\n- 时间复杂度:平均/最坏 O(n^2),最好 O(n)(当已基本有序且使用提前退出优化)\n- 空间复杂度:O(1)(就地排序)\n- 稳定性:稳定(相等元素相对次序不变)\n- 适用场景:数据量小或序列几乎有序;不适合大规模无序数据\n\n常见优化\n- 提前退出:某趟未发生交换则已排序完成。\n- 缩短边界:记录本趟最后一次交换的位置,下一趟只需遍历到该位置。\n- 双向冒泡(鸡尾酒排序):从左到右、再从右到左各走一遍,更快收敛。\n\n简要伪代码\nfunction bubbleSort(a):\n    n = length(a)\n    repeat\n        swapped = false\n        for j = 0 to n-2:\n            if a[j] > a[j+1]:\n                swap(a[j], a[j+1])\n                swapped = true\n        n = n - 1    // 末尾已就位,缩短范围\n    until swapped == false\n\n如需,我可以给出某种语言的实现代码或更多变体说明。",
        "refusal": null,
        "annotations": []
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 24,
    "completion_tokens": 1030,
    "total_tokens": 1054,
    "prompt_tokens_details": {
      "cached_tokens": 0,
      "audio_tokens": 0
    },
    "completion_tokens_details": {
      "reasoning_tokens": 512,
      "audio_tokens": 0,
      "accepted_prediction_tokens": 0,
      "rejected_prediction_tokens": 0
    }
  }
}
  • Unified chat API endpoint supporting all major text generation models
  • Fully compatible with the OpenAI Chat Completions API format
  • Seamlessly switch between different AI models via the model parameter
  • Supports streaming output, function calling, and other advanced features
curl -X POST https://qingbo.dev/v1/chat/completions \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "gpt-5",
    "messages": [
      {
        "role": "system",
        "content": "你是一个专业的AI助手。"
      },
      {
        "role": "user",
        "content": "解释一下冒泡排序算法。"
      }
    ],
    "temperature": 0.7,
    "max_tokens": 1024
  }'
{
  "id": "chatcmpl-CdA8u0qPqzeN2gu8QLuQfjaP9CHU4",
  "object": "chat.completion",
  "created": 1763449816,
  "model": "gpt-5-2025-08-07",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "冒泡排序(Bubble Sort)是一种简单的比较交换排序。它通过重复遍历序列,比较相邻元素并把较大的"冒泡"到右侧末端,直到序列有序。\n\n核心思想\n- 每一趟从左到右比较相邻元素,若前者大于后者则交换。一次趟结束后,当前未排部分中最大的元素被放到末尾。\n- 重复多趟,直到没有发生交换或已完成 n−1 趟。\n\n流程示例(以 [5, 1, 4, 2] 为例)\n- 第1趟:比较并交换 → [1, 5, 4, 2] → [1, 4, 5, 2] → [1, 4, 2, 5](最大元素5到末尾)\n- 第2趟:比较并交换 → [1, 4, 2, 5] → [1, 2, 4, 5](已整体有序,可提前结束)\n\n性质\n- 时间复杂度:平均/最坏 O(n^2),最好 O(n)(当已基本有序且使用提前退出优化)\n- 空间复杂度:O(1)(就地排序)\n- 稳定性:稳定(相等元素相对次序不变)\n- 适用场景:数据量小或序列几乎有序;不适合大规模无序数据\n\n常见优化\n- 提前退出:某趟未发生交换则已排序完成。\n- 缩短边界:记录本趟最后一次交换的位置,下一趟只需遍历到该位置。\n- 双向冒泡(鸡尾酒排序):从左到右、再从右到左各走一遍,更快收敛。\n\n简要伪代码\nfunction bubbleSort(a):\n    n = length(a)\n    repeat\n        swapped = false\n        for j = 0 to n-2:\n            if a[j] > a[j+1]:\n                swap(a[j], a[j+1])\n                swapped = true\n        n = n - 1    // 末尾已就位,缩短范围\n    until swapped == false\n\n如需,我可以给出某种语言的实现代码或更多变体说明。",
        "refusal": null,
        "annotations": []
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 24,
    "completion_tokens": 1030,
    "total_tokens": 1054,
    "prompt_tokens_details": {
      "cached_tokens": 0,
      "audio_tokens": 0
    },
    "completion_tokens_details": {
      "reasoning_tokens": 512,
      "audio_tokens": 0,
      "accepted_prediction_tokens": 0,
      "rejected_prediction_tokens": 0
    }
  }
}

Authorizations

Authorization
string
required
All endpoints require Bearer Token authenticationObtain an API Key:Visit the API Key management page to get your API KeyAdd the following to your request headers:
Authorization: Bearer YOUR_API_KEY

Body

model
string
required
Model nameSupported models include:
  • OpenAI: gpt-5, gpt-5-chat-latest, gpt-5-mini, gpt-5-nano, gpt-5-pro
  • Anthropic: claude-sonnet-4-5-20250929, claude-opus-4-1-20250805, claude-haiku-4-5-20251001, claude-opus-4-1-20250805-thinking, claude-sonnet-4-5-20250929-thinking
  • Google: gemini-2.5-pro, gemini-2.5-flash, gemini-2.5-pro-thinking, gemini-2.5-flash-lite,gemini-3-pro-preview-thinking,gemini-3-pro-image-preview,gemini-3-pro-preview
  • DeepSeek: deepseek-v3.1-250821, deepseek-v3.1-think-250821, deepseek-v3-0324
  • Doubao: doubao-seed-1-6-251015, doubao-seed-1-6-flash-250828, doubao-seed-1-6-thinking-250715
  • X.AI: grok-4.1
  • More models are continuously being added…
messages
array
required
Conversation message list
temperature
number
Controls output randomness, range 0-2
  • Lower values (e.g., 0.2) produce more deterministic output
  • Higher values (e.g., 1.8) produce more random output
Default: 1.0
max_tokens
integer
Maximum number of tokens to generateDifferent models have different maximum limits. Refer to the specific model documentation.
stream
boolean
Whether to enable streaming output
  • true: Stream response (SSE format)
  • false: Return complete response at once
Default: false
top_p
number
Nucleus sampling parameter, range 0-1Controls the diversity of generated text. It is recommended to use either this or temperature, not both.Default: 1.0
frequency_penalty
number
Frequency penalty, range -2.0 to 2.0Positive values reduce the likelihood of repeating the same tokensDefault: 0
presence_penalty
number
Presence penalty, range -2.0 to 2.0Positive values increase the likelihood of discussing new topicsDefault: 0
stop
string or array
Stop sequencesUp to 4 sequences. Generation stops when any of these sequences is encountered.
n
integer
Number of responses to generateDefault: 1

Response

id
string
Unique identifier for the response
object
string
Object type, always chat.completion
created
integer
Creation timestamp
model
string
The model name actually used
choices
array
List of generated responses
usage
object
Token usage statistics
system_fingerprint
string
System fingerprint identifier (for tracking backend configuration)

Supported Model List

OpenAI Series

  • gpt-5 - GPT-5 base model
  • gpt-5-chat-latest - GPT-5 latest chat version
  • gpt-5-mini - GPT-5 lightweight version, cost-effective
  • gpt-5-nano - GPT-5 ultra-lightweight version
  • gpt-5-pro - GPT-5 professional enhanced version
###X.ai Series
  • grok-4.1 - Grok 4.1 fast response version

Anthropic Series

  • claude-haiku-4-5-20251001 - Claude 4.5 fast response version
  • claude-sonnet-4-5-20250929 - Claude 4.5 balanced version
  • claude-opus-4-1-20250805 - Most powerful Claude 4.1 flagship model
  • claude-opus-4-1-20250805-thinking - Claude 4.1 Opus deep thinking version
  • claude-sonnet-4-5-20250929-thinking - Claude 4.5 Sonnet deep thinking version

Google Series

  • gemini-2.5-flash - Gemini 2.5 fast version
  • gemini-2.5-pro - Gemini 2.5 professional version
  • gemini-2.5-flash-lite - Gemini 2.5 ultra-lightweight version
  • gemini-2.5-pro-thinking - Gemini 2.5 Pro deep thinking version
  • gemini-3-pro-preview-thinking - Gemini 3 Pro deep thinking version
  • gemini-3-pro-preview - Gemini 3 Pro deep thinking version

DeepSeek Series

  • deepseek-v3.1-250821 - DeepSeek V3.1 base version
  • deepseek-v3.1-think-250821 - DeepSeek V3.1 thinking version
  • deepseek-v3-0324 - DeepSeek V3 standard version

Doubao Series

  • doubao-seed-1-6-flash-250828 - Doubao Seed 1.6 fast version
  • doubao-seed-1-6-thinking-250715 - Doubao Seed 1.6 thinking version
  • doubao-seed-1-6-251015 - Doubao Seed 1.6 standard version

Usage Examples

Basic Conversation

{
  "model": "gpt-5",
  "messages": [
    {"role": "user", "content": "你好"}
  ]
}

System Prompt

{
  "model": "claude-sonnet-4-5-20250929",
  "messages": [
    {"role": "system", "content": "你是一位专业的Python编程导师"},
    {"role": "user", "content": "如何使用列表推导式?"}
  ]
}

Multi-Turn Conversation

{
  "model": "gemini-2.5-flash",
  "messages": [
    {"role": "user", "content": "什么是机器学习?"},
    {"role": "assistant", "content": "机器学习是人工智能的一个分支..."},
    {"role": "user", "content": "能举个例子吗?"}
  ]
}

Streaming Output

{
  "model": "gpt-5",
  "messages": [
    {"role": "user", "content": "写一首关于春天的诗"}
  ],
  "stream": true
}

Advanced Features

This endpoint supports basic text chat. For the following advanced features, please refer to: