Top 5 This Week

Related Posts

Less is more: Meta study shows shorter reasoning improves AI accuracy by 34%


Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Researchers from Metaโ€™s FAIR team and The Hebrew University of Jerusalem have discovered that forcing large language models to โ€œthinkโ€ less actually improves their performance on complex reasoning tasks.

The study released today found that shorter reasoning processes in AI systems lead to more accurate results while significantly reducing computational costs.

โ€œIn this work, we challenge the assumption that long thinking chains results in better reasoning capabilities,โ€ write the authors in their paper titled โ€œDonโ€™t Overthink it. Preferring Shorter Thinking Chains for Improved LLM Reasoning.โ€

The research contradicts the prevailing trend in AI development, where companies have invested heavily in scaling up computing resources to allow models to perform extensive reasoning through lengthy โ€œthinking chainsโ€ โ€” detailed step-by-step trajectories that AI systems use to solve complex problems.

AI accuracy jumps 34% when models use shorter reasoning chains

The researchers discovered that within the same reasoning task, โ€œshorter reasoning chains are significantly more likely to yield correct answers โ€” up to 34.5% more accurate than the longest chain sampled for the same question.โ€ This finding held true across multiple leading AI models and benchmarks.

โ€œWhile demonstrating impressive results, [extensive reasoning] incurs significant computational costs and inference time,โ€ the authors note, pointing to a substantial inefficiency in how these systems are currently deployed.

Based on these findings, the team developed a novel approach called โ€œshort-m@k,โ€ which executes multiple reasoning attempts in parallel but halts computation once the first few processes complete. The final answer is then selected through majority voting among these shorter chains.

New โ€˜short-m@kโ€™ method slashes computing costs by 40% while boosting performance

For organizations deploying large AI reasoning systems, the implications could be substantial. The researchers found their method could reduce computational resources by up to 40% while maintaining the same level of performance as standard approaches.

โ€œShort-3@k, while slightly less efficient than short-1@k, consistently surpasses majority voting across all compute budgets, while still being substantially faster (up to 33% wall time reduction),โ€ the paper states.

Michael Hassid, the paperโ€™s lead author, and his team also discovered that training AI models on shorter reasoning examples improved their performance โ€” challenging another fundamental assumption in AI development.

โ€œTraining on the shorter ones leads to better performance,โ€ the researchers write. โ€œConversely, finetuning on S1-long increases reasoning time with no significant performance gains.โ€

Tech giants could save millions by implementing โ€œdonโ€™t overthink itโ€ approach

The findings come at a critical time for the AI industry, as companies race to deploy increasingly powerful models that consume enormous computational resources.

โ€œOur findings suggest rethinking current methods of test-time compute in reasoning LLMs, emphasizing that longer โ€˜thinkingโ€™ does not necessarily translate to improved performance and can, counter-intuitively, lead to degraded results,โ€ the researchers conclude.

โ€˜This research stands in contrast to other prominent approaches. Previous influential studies, including OpenAIโ€™s work on โ€œchain-of-thoughtโ€ prompting and โ€œself-consistencyโ€ methods, have generally advocated for more extensive reasoning processes. It also builds upon recent work like Princeton and Google DeepMindโ€™s โ€œTree of Thoughtsโ€ framework and Carnegie Mellonโ€™s โ€œSelf-Refineโ€ methodology, which have explored different approaches to AI reasoning.

For technical decision makers evaluating AI investments, the research suggests that bigger and more computationally intensive isnโ€™t always better. The study points toward potential cost savings and performance improvements by optimizing for efficiency rather than raw computing power.

In an industry obsessed with scaling up, it turns out that teaching AI to be more concise doesnโ€™t just save computing power โ€” it makes the machines smarter too. Sometimes, even artificial intelligence benefits from the age-old wisdom: donโ€™t overthink it.

#Meta #study #shows #shorter #reasoning #improves #accuracy
source: https://venturebeat.com/ai/less-is-more-meta-study-shows-shorter-reasoning-improves-ai-accuracy-by-34/

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Popular Articles