Skip to main content

Search site

Find podcasts, news, articles, webinars, and contributors in one search.

Study: Large Reasoning Models Demonstrate Limitations With Complex Tasks

Source: Apple Machine Learning

Found this useful? Share it with your network

Recent advancements in Large Reasoning Models (LRMs) reveal their capacity to generate detailed reasoning processes, yet their fundamental capabilities remain unclear. Current evaluations focus on final answer accuracy, neglecting the reasoning processes. This study uses controllable puzzle environments to analyze both final responses and internal reasoning traces. Findings indicate that while LRMs initially improve with increasing complexity, they ultimately experience a drop in accuracy, exhibiting a counterintuitive scaling behavior. In low complexity tasks, standard models outperform LRMs, while LRMs excel in medium complexity tasks, but both models struggle with high complexity, leading to significant performance collapse.

Read Full Article

Opens on Apple Machine Learning