POST
/
chat
/
completions
curl --request POST \
  --url https://api.siliconflow.cn/v1/chat/completions \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "model": "Qwen/QwQ-32B",
  "messages": [
    {
      "role": "user",
      "content": "What opportunities and challenges will the Chinese large model industry face in 2025?"
    }
  ],
  "stream": false,
  "max_tokens": 512,
  "stop": null,
  "temperature": 0.7,
  "top_p": 0.7,
  "top_k": 50,
  "frequency_penalty": 0.5,
  "n": 1,
  "response_format": {
    "type": "text"
  },
  "tools": [
    {
      "type": "function",
      "function": {
        "description": "<string>",
        "name": "<string>",
        "parameters": {},
        "strict": false
      }
    }
  ]
}'
{
  "id": "<string>",
  "choices": [
    {
      "message": {
        "role": "assistant",
        "content": "<string>",
        "reasoning_content": "<string>",
        "tool_calls": [
          {
            "id": "<string>",
            "type": "function",
            "function": {
              "name": "<string>",
              "arguments": "<string>"
            }
          }
        ]
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 123,
    "completion_tokens": 123,
    "total_tokens": 123
  },
  "created": 123,
  "model": "<string>",
  "object": "chat.completion"
}

Authorizations

Authorization
string
header
required

Use the following format for authentication: Bearer <your api key>

Body

application/json
model
enum<string>
default:Qwen/QwQ-32B
required

Corresponding Model Name. To better enhance service quality, we will make periodic changes to the models provided by this service, including but not limited to model on/offlining and adjustments to model service capabilities. We will notify you of such changes through appropriate means such as announcements or message pushes where feasible.

Available options:
Qwen/QwQ-32B,
Pro/deepseek-ai/DeepSeek-R1,
Pro/deepseek-ai/DeepSeek-V3,
deepseek-ai/DeepSeek-R1,
deepseek-ai/DeepSeek-V3,
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B,
deepseek-ai/DeepSeek-R1-Distill-Qwen-14B,
deepseek-ai/DeepSeek-R1-Distill-Qwen-7B,
deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B,
Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B,
Pro/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B,
deepseek-ai/DeepSeek-V2.5,
Qwen/Qwen2.5-72B-Instruct-128K,
Qwen/Qwen2.5-72B-Instruct,
Qwen/Qwen2.5-32B-Instruct,
Qwen/Qwen2.5-14B-Instruct,
Qwen/Qwen2.5-7B-Instruct,
Qwen/Qwen2.5-Coder-32B-Instruct,
Qwen/Qwen2.5-Coder-7B-Instruct,
Qwen/Qwen2-7B-Instruct,
Qwen/Qwen2-1.5B-Instruct,
Qwen/QwQ-32B-Preview,
TeleAI/TeleChat2,
THUDM/glm-4-9b-chat,
Vendor-A/Qwen/Qwen2.5-72B-Instruct,
internlm/internlm2_5-7b-chat,
internlm/internlm2_5-20b-chat,
Pro/Qwen/Qwen2.5-7B-Instruct,
Pro/Qwen/Qwen2-7B-Instruct,
Pro/Qwen/Qwen2-1.5B-Instruct,
Pro/THUDM/chatglm3-6b,
Pro/THUDM/glm-4-9b-chat
Example:

"Qwen/QwQ-32B"

messages
object[]
required

A list of messages comprising the conversation so far.

stream
boolean
default:false

If set, tokens are returned as Server-Sent Events as they are made available. Stream terminates with data: [DONE]

Example:

false

max_tokens
integer
default:512

The maximum number of tokens to generate.

Required range: 1 <= x <= 16384
Example:

512

stop

Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence.

Example:

null

temperature
number
default:0.7

Determines the degree of randomness in the response.

Example:

0.7

top_p
number
default:0.7

The top_p (nucleus) parameter is used to dynamically adjust the number of choices for each predicted token based on the cumulative probabilities.

Example:

0.7

top_k
number
default:50
Example:

50

frequency_penalty
number
default:0.5
Example:

0.5

n
integer
default:1

Number of generations to return

Example:

1

response_format
object

An object specifying the format that the model must output.

tools
object[]

A list of tools the model may call. Currently, only functions are supported as a tool. Use this to provide a list of functions the model may generate JSON inputs for. A max of 128 functions are supported.

Response

200
application/json
200
id
string
choices
object[]
usage
object
created
integer
model
string
object
enum<string>
Available options:
chat.completion