Microsoft I launched many new “open” artificial intelligence models On Wednesday, what is able to compete with Openai’s O3-MINI at least one match.
All new licensed models-PHI 4 Mini logic, logic 4, and PLUS-Phi 4 Plus-are “thinking” models, which means that they are able to eliminate more fact examination solutions to complex problems. They are saving the Microsoft Phi “Small Model” family, which the company launched a year ago to provide a basis for AI developers for applications on the edge.
Phi 4 MINI logic has been trained in about one million artificial mathematics problems caused by the Deepseek model of artificial intelligence. Microsoft says about 3.8 billion teachers in size, PHI 4 Mini Reasoning for educational applications, such as “compact lessons” on light devices, says.
Parameters are almost compatible with problem solving skills in the model, and models with parameters generally perform better than those that contain less parameters.
Phi 4 Reasoning, a model of 14 billion teachers, was trained using “high-quality” web data in addition to “coordinated demonstrations” from the above O3-MINI from Openai. It is better for mathematics, science and coding applications, according to Microsoft.
For Phi 4 Indudeing Plus, the Microsoft model was pre-adapted to the thinking model to achieve better accuracy in certain tasks. Microsoft claims that Phi 4 Reasoning Plus is close to R1 performance levels, a model that contains much more parameters (671 billion). The company’s internal standard also contains Phi 4 Reasoning in addition to the O3-MINI matching on omnimath, which is a test of mathematics skills.
Thinking about MINI Phi 4, Phi 4, and Phi 4 Reasoning Plus on Dev Ai Embrace the face It is accompanied by detailed technical reports.
TECHRUNCH event
Berkeley, California
|
June 5
“Using distillation, reinforcement learning, high -quality data, these (new) models balance size and performance,” Microsoft wrote in A. Blog post. “It is small enough for low total environments, but it maintains strong thinking capabilities that compete with much larger models. This mixture allows even limited resources to perform complex thinking tasks efficiently.”
https://techcrunch.com/wp-content/uploads/2024/10/GettyImages-1883327378-e1730136121848.jpg?resize=1200,800
Source link