Join our daily and weekly newsletters for the latest updates and the exclusive content on AI coverage. Learn more
Brain systems Announced today that it will welcome the breakthrough of Deepseek R1 Artificial intelligence model on American serverspromising speeds up to 57 times faster than GPU -based solutions while retaining sensitive data inside American borders. This decision comes in the midst of growing concerns concerning the rapid AI progress and data confidentiality.
The Puce AI startup will deploy a version of 70 billion parameters of Deepseek-R1 By operating on its owner equipment at the scale of the brochure, offering 1,600 tokens per second – a spectacular improvement compared to traditional GPU implementations which fought with more recent “reasoning” models.
Why the Deepseek reasoning models reshape the company AI
“These reasoning models affect the economy,” said James Wang, a Cerebras senior manager in an exclusive interview with Venturebeat. “Any knowledge worker must mainly perform a kind of cognitive tasks in several stages. And these reasoning models will be the tools that enter their workflow. »»
The announcement follows a tumultuous week in which the emergence of Deepseek sparked Nvidia Most important loss of market value, nearly $ 600 billionLifting questions about the supremacy of the chip giant AI. The Cerebras solution responds directly to two key concerns that have emerged: the requirements for calculating advanced AI models and data sovereignty.
“If you use Deepseek APIWho is very popular at the moment, this data is sent directly to China, ”said Wang. “It is a strong warning that (fact) many American companies and companies … not willing to consider it (this one).”
How the technology of the Cerebras brochure beats traditional GPUs at AI speed
Cerebras obtains its advantage at speed thanks to a new chip architecture which retains whole AI models on a processor the size of a single brochure, eliminating the bottlenecks of memory that afflict GPU -based systems. The company claims that its implementation of Deepseek-R1 matches or exceeds the performance of OPENAI proprietary models, while fully operating on American soil.
Development represents a significant change in the AI landscape. In depthFounded by the former director of hedge funds, Liang Wenfeng, shocked the industry by reaching the reasoning capacities of sophisticated AI which would have been only 1% of the cost of American competitors. Cerebras’ accommodation solution now offers American companies a way to take advantage of these advances while retaining data control.
“It is actually a great story that American research laboratories have offered this gift to the world. The Chinese have taken it and improved it, but it has limits because it works in China, has censorship problems, and now we take them back and execute them on American data centers, without censorship, without retention of retention data, “said Wang.
American technological leadership faces new questions as IA innovation becomes global
The service will be available via a Developer overview From today. Although it is initially free, Cerebras plans to implement API access controls Due to high early demand.
This decision comes as American legislators face the implications of the rise of Deepseek, which has exposed potential limits US trade restrictions Designed to maintain technological advantages compared to China. The capacity of Chinese companies to achieve the perceneous capabilities of AI despite Chip export checks has aroused calls for new regulatory approaches.
Industry analysts suggest that this development could accelerate the discrepancy of the IA infrastructure dependent on the GPU. “Nvidia is no longer the leader in inference performance,” noted Wang, pointing benchmarks showing higher performance of various specialized AI chips. “These other IA flea companies are really faster than GPUs to manage these latest models.”
The impact extends beyond technical measures. As the AI models are increasingly incorporating sophisticated reasoning capacities, their calculation requests have skyrocketed. Cerebras maintains that its architecture is better suited to these emerging workloads, potentially reshaping the competitive landscape of the deployment of corporate AI.