Join daily and weekly newsletters to obtain the latest updates and exclusive content to cover the leading artificial intelligence in the industry. Learn more
Meta researchers Fair team and Hebrew University of Jerusalem They discovered that forcing large language models to “think” is less actually improves their performance in complex thinking tasks.
the Ticket Today I found that shorter thinking processes in artificial intelligence systems lead to more accurate results while significantly reducing calculations.
“In this work, we challenge the assumption that long thinking chains lead to better capabilities of thinking,” the authors write in their paper entitled.Do not think about it. Favorite chains of short thinking to improve thinking in llm“
The research contrasts with the prevailing trend in developing artificial intelligence, as companies have invested extensively in increasing computing resources to allow materials to perform widespread thinking through lengthy.Thinking chains-Step-step-by-step paths used by artificial intelligence systems to solve complex problems.
The resolution of artificial intelligence jumps by 34 % when models use shorter thinking chains
The researchers discovered that within the same task of thinking, “Luxor thinking chains are likely to resulted in the right answers – up to 34.5 % more accurate than the longest series of samples of them for the same question.” This discovery is true through several pioneering models and standards of Amnesty International.
“While showing impressive results, (broad thinking) bears great mathematical costs and the time of reasoning,” the authors note, noting a great inefficiency in how these systems are currently publishing.
Based on these results, the team developed a new approach called “Short m@k“Who carries out multiple thinking attempts in parallel, but stops the account as soon as the first few operations are completed. Then the final answer is chosen through the majority vote between these shorter chains.
New method ‘Short-m@K’ transmits computing costs by 40 % with increased performance
For institutions that publish large male thinking systems, the effects can be large. The researchers found that their method can reduce arithmetic resources by up to 40 % while maintaining the same level as standard methods.
“Short 3@K, although it is slightly less efficient than short 1@K, the majority voting constantly exceeds all mathematical budgets, while it is still largely faster (up to 33 % of the wall time),” says the paper.
Michael Hasid, the main author of the paper, and his team discovered that training artificial intelligence models on examples of shorter thinking improves its performance – another basic assumption challenge in developing artificial intelligence.
The researchers write: “Luxor training leads to better performance,” the researchers writes. “On the contrary, thyroiditis on the S1 increases the time of thinking with the lack of significant gains in performance.”
Technology giants can save millions by implementing a “do not think about it” approach
The results come in a crucial time for the artificial intelligence industry, as companies are racing to spread strong models that consume huge mathematical resources.
“The results we have reached referring to rethinking the methods of calculating the test time in thinking about LLMS, while emphasizing that the longer” thinking “does not necessarily translate to improved performance and can, inappropriately, lead to deteriorating results.
This research contradicts the opposite of other prominent methods. Previous impressive studies, including Openai’s work on “Thought” claim And “and”Self -consistency“The methods, generally called for the most comprehensive thinking. It also depends on the last work like Princeton and Google Deebind.”Ideas“Frame and Carnegie Mellon”A self -reverseThe methodology, which explored different approaches to the logic of artificial intelligence.
For technical decision makers who evaluate artificial intelligence investments, the research indicates that the largest and more calculation is not always better. The study indicates possible cost savings and improving performance by improving efficiency rather than raw computing strength.
In an expanded industry, it turns out that teaching artificial intelligence is more brief not only to provide computing power – it makes machines more intelligent. Sometimes, even artificial intelligence benefits from ancient wisdom: don’t think about it.
https://venturebeat.com/wp-content/uploads/2025/05/nuneybits_Vector_art_of_a_robot_thinking_deeply_image_in_Facebo_22675f0a-e908-4a6a-9502-56addbde1374.webp?w=1024?w=1200&strip=all
Source link