News

The Hangzhou-based firm said later in a short post on X that R1-0528 featured improved performance. In a longer post on WeChat, DeepSeek said the rate of "hallucinations", false or misleading ...
For instance, in the AIME 2025 test, DeepSeek-R1-0528’s accuracy jumped from 70% to 87.5%, indicating deeper reasoning processes that now average 23,000 tokens per question compared to 12,000 in ...
SHANGHAI/BEIJING -Chinese artificial intelligence startup DeepSeek released the first update to its hit R1 reasoning model in the early hours of Thursday, stepping up competition with U.S. rivals such ...
The updated R1 model, named R1-0528, was released quietly on the Hugging Face platform without an official announcement, suggesting DeepSeek’s preference for action over publicity.
DeepSeek also said it distilled the reasoning steps used in R1-0528 into Alibaba’s Qwen3 8B Base model. That process created a new, smaller model that surpassed Qwen3’s performance by more ...
DeepSeek launched R1-0528 on developer platform Hugging Face, but has yet to make an official public announcement. It did not publish a description of the model or comparisons.
DeepSeek launched R1-0528 on developer platform Hugging Face, but has yet to make an official public announcement. It did not publish a description of the model or comparisons.
DeepSeek-R1-0528's ranking (August 1, 2024 to May 1, 2025) is fourth at the time of writing, with a score showing performance almost on par with OpenAI's o4-mini (medium).
Despite the significant attention the R1 model garnered at its launch, the latest update was released with fewer details. However; DeepSeek later disclosed on X that the R1-0528 version boasted ...
China's free-for-all AI models, developed by firms like DeepSeek and Alibaba, present a viable alternative to US ...