Meta

Photo of author

By [email protected]


Join daily and weekly newsletters to obtain the latest updates and exclusive content to cover the leading artificial intelligence in the industry. Learn more


Dead Today I announced a partnership with Brain systems For its new operation Lama APIAllowing developers to reach up to 18 times faster than traditional GPU.

The advertisement, which was made in the opening dead For the cousins The Menlo Park developer conference, puts the company to compete directly with Openaiand manAnd Google In the rapidly growing inference services market, developers buy symbols by billions to run their applications.

“META chose the brain to cooperate to connect the very rapid reasoning they need to serve developers through their new Lamaa applications interface,” Julie Shen Choi, chief marketing official in Kerberra, said at a press conference. “We are really in Kerbras, are really excited to announce the first partnership of CSP Hyperscale to provide a quick reasoning for all developers.”

The partnership represents the official Meta entry into the work of selling the artificial intelligence account, which transforms the famous open source Lama models into a commercial service. While the Llama models accumulated from Meta One billion downloadSo far, the company has not provided the cloud infrastructure for the first party for developers to build applications with them.

“This is very exciting, even without talking about the brain specifically,” said James Wang, chief executive of Kerbras. “Openai, Anthropic, Google – they have built a completely new AI’s business from the zero point, which is artificial intelligence inference. Developers who are building AI applications will buy codes in millions, sometimes by billions. These are similar to the new account instructions that people need to build AI applications.”

The standard graph shows the Lama 4 brain processing at 2648 symbols per second, which greatly outperforms Sambanova (747), GROQ (600) and Google GPU services and others-Explanation of the choice of Meta devices for a new application programming interface. (Credit: Brain)

Break the speed barrier: How are the super -brain models

What distinguishes the Meta width from each other is the increase in the dramatic speed provided by the artificial intelligence chips specialized in Cerebras. The brain system delivers 2600 symbols per second For Llama 4 Scout, compared to about 130 icons per second for ChatGPT and about 25 icons per second for Deepseek, according to the criteria from Artificial analysis.

“If you only compare the API-TO-API, Gemini and GPT basis, they are all great models, but they all work at GPU speeds, which are about 100 code per second,” Wang explained. “And 100 code per second is fine to chat, but it is very slow for thinking. It’s very slow for agents. People are struggling with that day.”

This speed feature provides completely new categories of applications that were previously impractical, including factors in actual time, low audio systems to join the conversation, generate the interactive code, and immediate multi-step thinking-and all require a sequence of multiple linguistic calls that can be completed now in seconds instead of minutes.

the Lama API It represents a major shift in the Meta strategy of artificial intelligence, as it moves from being a typical provider until it becomes a fully -service artificial intelligence infrastructure company. By providing API service, Meta creates a flow of revenue from its artificial intelligence investments while maintaining its commitment to open the models.

“Meta is now in the field of selling distinctive symbols, which is great for the American ecosystems of artificial intelligence,” Wang said during the press conference. “They bring a lot to the table.”

The application programming interface will provide tools for control and evaluation, starting with Llama 3.3 8BAllow developers to create and train data and test the quality of their custom models. Meta emphasizes that it will not use customer data to train its own models, and the models designed using API Llama can be transferred to two other hosts – a clear distinction from the most closed methods of some competitors.

The brain will introduce the new Meta service through its network Data centers It is located throughout North America, including facilities in Dallas, Aklahoma, Minnesota, Montreal and California.

“All data centers that serve the reasoning are present in North America at this time,” explained by Choi. “We will provide dead with the full ability of the brain. The work burden will be balanced across all these different data centers.”

The work ranking follows what Choi described as “the classic account provider of the excessive performance model”, similar to how NVIDIA has devices for the main cloud providers. She said, “They keep our account blocs, which can serve the residents of the developers.”

Beyond the brain, Meta has also announced a partnership with Groq To provide rapid reasoning options, giving developers multiple high -performance alternatives exceeds the traditional GPU.

The entry of Meta to the API market can be disrupted with the outstanding performance standards of the applicable arrangement Openaiand GoogleAnd man. By combining the popularity of its open source models and the capabilities of great conclusion, META defines itself as a huge competitor in the area of ​​commercial artificial intelligence.

“Meta is in a unique position with 3 billion users, a severe data center, and an ecosystem for the huge developer,” according to the presentation of CEREBRAS. Merging brain technology “Meta Leapfrog Openai and Google help in about 20x.”

For the brain, this partnership represents a major milestone and verifying the authenticity of the specialized artificial intelligence devices. “I built this engine on the scale of biscuits for years, and we always knew that the first benefit of technology, but in the end it should end as part of another person’s cloud. This was the ultimate goal of the trade strategy perspective, and we finally reached this teacher,” Wang said.

the Lama API It is currently available as a limited inspection, with planning for a broader consulting in the coming weeks and months. Developers interested in accessing Llama 4 ultra -speed reasoning can ask for early access by choosing the brain from the model options within the API Llama.

“If you imagine a developer who does not know anything about the brain because we are a relatively small company, they can only click two button on the standard Meta SDK, create the API key, choose the brain mark, then suddenly, the symbols are processed on a giant chips.” Wang explained. “This type of ultimately made us the background of the Meta’s full developer’s ecosystem is great for us.”

Meta choosing a specialized silicone indicates a deep thing: in the next stage of artificial intelligence, not only what your models know, but how quickly you think about it. In this future, the speed is not just a feature – it’s the main point.



https://venturebeat.com/wp-content/uploads/2025/04/nuneybits_Vector_art_of_a_computer_chip_in_the_colors_of_Goolge_71db7087-b8f7-4ae3-ae07-4b36bdee2086.webp?w=1024?w=1200&strip=all
Source link

Leave a Comment