Release notes
Platform Service Adjustment Notice
To ensure the quality of platform services and the rational allocation of resources, the following adjustments to Rate Limits policies are now in effect:
- Adjustments
New RPH Limit (Requests Per Hour, Per Hour Requests)
- Model Scope:deepseek-ai/DeepSeek-R1, deepseek-ai/DeepSeek-V3
- Applicable Users: All users
- Limit Standard: 30 requests/hour
2.New RPD Limit (Requests Per Day, Per Day Requests)
- Model Scope: deepseek-ai/DeepSeek-R1, deepseek-ai/DeepSeek-V3
- Applicable Users: Users who have not completed real-name authentication
- Limit Standard: 100 requests/day
Please note that these policies may be adjusted at any time based on traffic and load changes. Silicon Flowing Reserves the right to interpret these policies.
1. Model Offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on February 27, 2025:
- meta-llama/Meta-Llama-3.1-405B-Instruct
- google/gemma-2-27b-it
- google/gemma-2-9b-it
- Pro/google/gemma-2-9b-it
- 01-ai/Yi-1.5-34B-Chat-16K
- 01-ai/Yi-1.5-6B-Chat
- 01-ai/Yi-1.5-9B-Chat-16K
- stabilityai/stable-diffusion-3-medium
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
DeepSeek-V3 model prices have been restored to the original price starting from Beijing time February 9, 2025, at 00:00.
Specific prices:
- Input: ¥2/M Tokens
- Output: ¥8/M Tokens
Inference model output adjustment notice
The display of the reasoning chain in the inference model will be separated into a separate reasoning_content field from the content field. This change is compatible with the OpenAI and DeepSeek API specifications, making it easier for various frameworks and upper-layer applications to trim the conversation in multi-round dialogues. For more details, see the Inference Model (DeepSeek-R1) Usage.
Platform service adjustment notice
Support for deepseek-ai/DeepSeek-R1 and deepseek-ai/DeepSeek-V3 Models
The specific pricing is as follows:
deepseek-ai/DeepSeek-R1
Input:¥4/ M Tokens Output: ¥16/ M Tokensdeepseek-ai/DeepSeek-V3
- From February 1, 2025, to February 8, 2025, 24:00 Beijing Time, enjoy a limited-time discount price:Input:
¥2¥1/ M Tokens Output:¥8¥2/ M Tokens,The original price will be restored from February 9, 2025, 00:00.
- From February 1, 2025, to February 8, 2025, 24:00 Beijing Time, enjoy a limited-time discount price:Input:
Platform service adjustment notice
Image and Video URL Validity Period Adjusted to 1 Hour
To continue providing you with more advanced and high-quality services, the validity period of image and video URLs generated by large models will be adjusted to 1 hour starting from January 20, 2025.
If you are currently using the image and video generation service, please make sure to back up the files in time to avoid any business disruptions due to URL expiration.
Platform service adjustment notice
LTX-Video Model Will Start Charging
To continue providing you with more advanced and high-quality services, the platform will start charging for video generation requests using the Lightricks/LTX-Video model starting from January 6, 2025, at a rate of 0.14 yuan per video.
Platform service adjustment notice
1.New Global Access API Endpoint Added
A new global access API endpoint has been added: https://api.siliconflow.com. If you encounter network connection issues while using the original endpoint https://api.siliconflow.cn, it is recommended to switch to the new endpoint.
Platform service adjustment notice
1. Model Offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on December 19, 2024:
- deepseek-ai/DeepSeek-V2-Chat
- Qwen/Qwen2-72B-Instruct
- Vendor-A/Qwen/Qwen2-72B-Instruct
- OpenGVLab/InternVL2-Llama3-76B
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
1. Model offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on December 13, 2024:
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
1. Model offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on November 22, 2024:
- deepseek-ai/DeepSeek-Coder-V2-Instruct
- meta-llama/Meta-Llama-3-70B-Instruct
- meta-llama/Meta-Llama-3-8B-Instruct
- Qwen/Qwen2-57B-A14B-Instruct
- Pro/internlm/internlm2_5-7b-chat
- Pro/meta-llama/Meta-Llama-3-8B-Instruct
- Pro/THUDM/chatglm3-6b
- Pro/01-ai/Yi-1.5-9B-Chat-16K
- Pro/01-ai/Yi-1.5-6B-Chat
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
2.Email login method update
To further enhance service experience, the platform will update the login method starting from November 22, 2024: from the original “email account + password” method to an “email account + verification code” method.
3. New Overseas API Endpoint
A new endpoint for overseas users has been added: https://api-st.siliconflow.cn. If you encounter network connection issues while using the original endpoint https://api.siliconflow.cn, it is recommended to switch to the new endpoint.
Partial Model Pricing Adjustment Notice
To provide more stable, high-quality, and sustainable services, the Vendor-A/Qwen/Qwen2-72B-Instruct model, which was previously offered for free, will start charging from October 17, 2024. The pricing details are as follows:
- Limited-time discount price:¥ 1.00 / M tokens
- Original price:¥ 4.13 / M tokens(the original price will be restored at a later date)