Release notes
Platform Service Adjustment Notice
As of now, the Pro/deepseek-ai/DeepSeek-V3
and deepseek-ai/DeepSeek-V3
model has been updated to the latest 0324 version. You can still use the older model via Pro/deepseek-ai/DeepSeek-V3-1226
to facilitate a smoother transition of your business.
Platform Service Adjustment Notice
SiliconCloud will initiate the update of the DeepSeek V3 model.
For the deepseek-ai/DeepSeek-V3
and Pro/deepseek-ai/DeepSeek-V3
models, they will be “progressively” updated to the latest 0324 version.
After the update, both models will be at the 0324 version. If needed, you can still use the old version of the model via deepseek-ai/DeepSeek-V3-1226
until April 30, 2025, to facilitate a smoother business transition.
Platform Service Adjustment Notice
To better serve global developer users, SiliconCloud will soon launch an international site and gradually open multiple service regions.
Due to this adjustment, the existing api.siliconflow.com API endpoint will be phased out at an appropriate time. Please switch to api.siliconflow.cn as soon as possible to continue using the service.
We have already configured Global Traffic Manager (GTM) for the .cn endpoint to provide the same global access experience as the current .com endpoint. You only need to change the base URL of your API requests to api.siliconflow.cn.
We recommend that you complete the migration by the end of this month (March 31). If you have any questions, please contact us at any time.
Platform Service Adjustment Notice
To continuously improve user experience, the Rate Limits policy is being adjusted as follows:
Remove the RPH and RPD rate limits for deepseek-ai/DeepSeek-R1 and deepseek-ai/DeepSeek-V3.
As traffic and load change, the policy may be adjusted at any time, and Silicic Flow reserves the right to interpret.
Platform Service Adjustment Notice
1. Model Offline notice
To further optimize resource allocation and provide more advanced, high-quality, and compliant technical services, the platform will shut down certain models on March 6, 2025. The specific list of models involved is as follows:
- Chat models:
- AIDC-AI/Marco-o1
- meta-llama/Meta-Llama-3.1-8B-Instruct
- Pro/meta-llama/Meta-Llama-3.1-8B-Instruct
- meta-llama/Meta-Llama-3.1-70B-Instruct
- meta-llama/Meta-Llama-3.1-405B-Instruct
- meta-llama/Llama-3.3-70B-Instruct
- Image generation models:
- black-forest-labs/FLUX.1-schnell
- Pro/black-forest-labs/FLUX.1-schnell
- black-forest-labs/FLUX.1-dev
- black-forest-labs/FLUX.1-pro
- stabilityai/stable-diffusion-xl-base-1.0
- stabilityai/stable-diffusion-3-5-large
- stabilityai/stable-diffusion-3-5-large-turbo
- stabilityai/stable-diffusion-2-1
- deepseek-ai/Janus-Pro-7B
- Voice models:
- fishaudio/fish-speech-1.5
- FunAudioLLM/SenseVoiceSmall
- FunAudioLLM/CosyVoice2-0.5B
- fishaudio/fish-speech-1.4
- RVC-Boss/GPT-SoVITS
- Video models:
- Lightricks/LTX-Video
- genmo/mochi-1-preview
Platform Service Adjustment Notice
To ensure the quality of platform services and the rational allocation of resources, the following adjustments to Rate Limits policies are now in effect:
- Adjustments
New RPH Limit (Requests Per Hour, Per Hour Requests)
- Model Scope:deepseek-ai/DeepSeek-R1, deepseek-ai/DeepSeek-V3
- Applicable Users: All users
- Limit Standard: 30 requests/hour
2.New RPD Limit (Requests Per Day, Per Day Requests)
- Model Scope: deepseek-ai/DeepSeek-R1, deepseek-ai/DeepSeek-V3
- Applicable Users: Users who have not completed real-name authentication
- Limit Standard: 100 requests/day
Please note that these policies may be adjusted at any time based on traffic and load changes. Silicon Flowing Reserves the right to interpret these policies.
1. Model Offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on February 27, 2025:
- 01-ai/Yi-1.5-34B-Chat-16K
- 01-ai/Yi-1.5-6B-Chat
- 01-ai/Yi-1.5-9B-Chat-16K
- stabilityai/stable-diffusion-3-medium
- google/gemma-2-27b-it
- google/gemma-2-9b-it
- Pro/google/gemma-2-9b-it
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
DeepSeek-V3 model prices have been restored to the original price starting from Beijing time February 9, 2025, at 00:00.
Specific prices:
- Input: ¥2/M Tokens
- Output: ¥8/M Tokens
Inference model output adjustment notice
The display of the reasoning chain in the inference model will be separated into a separate reasoning_content field from the content field. This change is compatible with the OpenAI and DeepSeek API specifications, making it easier for various frameworks and upper-layer applications to trim the conversation in multi-round dialogues. For more details, see the Inference Model (DeepSeek-R1) Usage.
Platform service adjustment notice
Support for deepseek-ai/DeepSeek-R1 and deepseek-ai/DeepSeek-V3 Models
The specific pricing is as follows:
deepseek-ai/DeepSeek-R1
Input:¥4/ M Tokens Output: ¥16/ M Tokensdeepseek-ai/DeepSeek-V3
- From February 1, 2025, to February 8, 2025, 24:00 Beijing Time, enjoy a limited-time discount price:Input:
¥2¥1/ M Tokens Output:¥8¥2/ M Tokens,The original price will be restored from February 9, 2025, 00:00.
- From February 1, 2025, to February 8, 2025, 24:00 Beijing Time, enjoy a limited-time discount price:Input:
Platform service adjustment notice
Image and Video URL Validity Period Adjusted to 1 Hour
To continue providing you with more advanced and high-quality services, the validity period of image and video URLs generated by large models will be adjusted to 1 hour starting from January 20, 2025.
If you are currently using the image and video generation service, please make sure to back up the files in time to avoid any business disruptions due to URL expiration.
Platform service adjustment notice
LTX-Video Model Will Start Charging
To continue providing you with more advanced and high-quality services, the platform will start charging for video generation requests using the Lightricks/LTX-Video model starting from January 6, 2025, at a rate of 0.14 yuan per video.
Platform service adjustment notice
1. Model Offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on December 19, 2024:
- deepseek-ai/DeepSeek-V2-Chat
- Qwen/Qwen2-72B-Instruct
- Vendor-A/Qwen/Qwen2-72B-Instruct
- OpenGVLab/InternVL2-Llama3-76B
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
1. Model offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on December 13, 2024:
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
Platform service adjustment notice
1. Model offline notice
To provide more stable, high-quality, and sustainable services, the following models will be offline on November 22, 2024:
- deepseek-ai/DeepSeek-Coder-V2-Instruct
- Qwen/Qwen2-57B-A14B-Instruct
- Pro/internlm/internlm2_5-7b-chat
- Pro/THUDM/chatglm3-6b
- Pro/01-ai/Yi-1.5-9B-Chat-16K
- Pro/01-ai/Yi-1.5-6B-Chat
If you are using any of these models, it is recommended to migrate to other models available on the platform as soon as possible.
2.Email login method update
To further enhance service experience, the platform will update the login method starting from November 22, 2024: from the original “email account + password” method to an “email account + verification code” method.
3. New Overseas API Endpoint
A new endpoint for overseas users has been added: https://api-st.siliconflow.cn. If you encounter network connection issues while using the original endpoint https://api.siliconflow.cn, it is recommended to switch to the new endpoint.
Partial Model Pricing Adjustment Notice
To provide more stable, high-quality, and sustainable services, the Vendor-A/Qwen/Qwen2-72B-Instruct model, which was previously offered for free, will start charging from October 17, 2024. The pricing details are as follows:
- Limited-time discount price:¥ 1.00 / M tokens
- Original price:¥ 4.13 / M tokens(the original price will be restored at a later date)