▶Recursive models like HRM and TRM demonstrate superior performance over large language models on specific, complex reasoning benchmarks such as ArcPrize 1.May 2026
▶Standard feed-forward transformer architectures have inherent limitations, making them unable to solve problems that require more sequential steps than the number of layers in the model.May 2026
▶Model efficiency is a key advantage of recursive architectures, with models as small as 7 million parameters outperforming trillion-parameter LLMs on certain tasks.May 2026
▶Innovations in training methods, such as DEQ-inspired techniques in HRM and full backpropagation through latent steps in TRM, have been critical to the success of recursive reasoning models.May 2026
▶While acknowledging that LLMs are Turing-complete at test time with methods like Chain of Thought, Chauvard argues they are fundamentally incapable of solving certain problems that smaller recursive models can, presenting a nuanced view on LLM capabilities.May 2026
▶There is a clear evolution in architectural approach, from the three-loop, dual-network Hierarchical Reasoning Model (HRM) to the more streamlined and efficient single, weight-shared network of the Tiny Recursive Model (TRM).May 2026
▶Chauvard contrasts the historical failure of backpropagation through time in RNNs due to error accumulation with the recent success of new backpropagation strategies in HRM and TRM, which have been key to their performance.May 2026
▶Within the HRM architecture, Chauvard highlights that analysis by a colleague identified the 'outer refinement loop' as the primary driver of its high performance, suggesting an internal discovery process about which components were most critical.May 2026
Not enough data for timeline
Sign up free to see the full intelligence report
Get started free