Cerebrra’s become the world’s fastest host for Deepseek R1, who exceeds Nvidia GPUs with 57x

Cerebrra's become the world's fastest host for Deepseek R1, who exceeds Nvidia GPUs with 57x

Become a member of our daily and weekly newsletters for the latest updates and exclusive content about leading AI coverage. Leather


Cerebras systems Announced today that it will organize the breakthrough of Deepseek R1 Artificial intelligence model on American serversPromising speeds up to 57 times faster than GPU -based solutions, while sensitive data is stored within American boundaries. The move comes in the midst of growing concerns about the fast AI preface and data privacy of China.

The AI ​​Chip Startup will implement a 70 billion parameter version of Deepseek-R1 Run on its own hardware on wafer scale, which provides 1,600 tokens per second-a dramatic improvement compared to traditional GPU implementations that have struggled with newer “reasoning” AI models.

Reaction times of leading AI platforms, measured in seconds. Cerebras reaches the fastest response to just over a second, while the Novita system takes almost 38 seconds to generate its first output-a critical metric for applications in practice. (Source: artificial analysis)

Why reform the Reasoning models of Deepseek de Enterprise AI

“These reasoning models have an influence on the economy,” said James Wang, a senior executive at Cerebrra’s, in an exclusive interview with Venturebeat. “In principle, every knowledge worker must perform a kind of cognitive tasks with multiple steps. And these reasoning models will be the tools that introduce their workflow. “

The announcement follows a tumultuous week in which the rise of Deepseek has caused Nvidia The biggest loss of market value ever, almost $ 600 billionCalling questions about the AI ​​supremation of the chip giant. Cerebras’s solution is immediately engaged in tackling two important concerns that have emerged: the calculation requirements of advanced AI models and data display.

See also  OnePlus 13 Releasedatum, prijs en specificaties Geruchten

‘If you use it Deepseek’s APIWhat is now very popular that data is sent directly to China, “Wang explained. “That is a serious reservation that [makes] Many American companies and companies … not willing to consider [it]. “

CereBra’s demonstrates dramatic performance benefits in the output speed and process 1,508 tokens per second-bine six times faster than the closest competitor, groq and about 100 times faster than traditional GPU-based solutions such as Novita. (Source: artificial analysis)

How Cerebras’s wafer scaling technology beats the traditional GPUs with AI speed

Cerebras reaches its speed benefit due to a new chip architecture that holds entire AI models in a single wafer format processor, which eliminates the based on memory based on GPU-based systems. The company claims the implementation of Deepseek-R1 competitions or surpasses the performance of OpenAi’s own models, while it runs entirely on American soil.

The development represents an important shift in the AI ​​landscape. DeepFounded by former Hedgefonds director Liang Wenfeng, the industry shocked by advanced AI reasoning possibilities to only only 1% of the costs of American competitors. Cerebras’s hosting solution now offers American companies a way to make this progress while retaining data control.

“It is actually a nice story that the US Research Labs have given this gift to the world. The Chinese took it and improved it, but it has limitations because it is in China, has some censorship problems, and now we take it back and perform it on American data centers, without censorship, without data retention, “Wang said.

Performance benchmarks with Deepseek-R1 that run on cerebras, perform better than both GPT-4O and OpenAI’s O1-Mini about questions that answer, mathematical reasoning and coding tasks. The results suggest that Chinese AI development may approach or exceed the American possibilities in some areas. (Credit: cerebras)

US Tech Leadership is confronted with new questions, because AI innovation goes worldwide

The service is available through a Developer Preview From today. Although it will initially be free, Cerebras is planning to implement API Access Controls Because of a strong early question.

See also  'No, No, No!': CNN Host Hits Back At Rep. Byron Donalds Over Trump's Women Comment

The move comes as the American legislators struggle with the implications of Deepseek’s rise, which has exposed potential restrictions in American commercial restrictions Designed to maintain technological benefits compared to China. The ability of Chinese companies to achieve breakthrough AI possibilities, despite Chip -Export checks has requested calls for new legal approaches.

Industrial analysts suggest that this development could accelerate the shift of GPU-dependent AI infrastructure. “Nvidia is no longer the leader in the conclusion,” Wang noted, pointing to benchmarks that show superior performance of various specialized AI chips. “These other AI chip companies are really faster than GPUs for performing these latest models.”

The impact extends beyond technical statistics. Because AI models are increasingly taking advanced reasoning opportunities, their computational requirements have been raised. Cerebras claims that his architecture is better suited for this emerging workload, which may reform the competitive landscape in the implementation of Enterprise AI.


Source link