ArXivIQ

ArXivIQ

Share this post

ArXivIQ
ArXivIQ
Hierarchical Reasoning Model

Hierarchical Reasoning Model

Aug 02, 2025
∙ Paid

Share this post

ArXivIQ
ArXivIQ
Hierarchical Reasoning Model
Share

Authors: Guan Wang, Jin Li, Yuhao Sun, Xing Chen, Changling Liu, Yue Wu, Meng Lu, Sen Song, Yasin Abbasi Yadkori
Paper: https://arxiv.org/abs/2506.21734
Code: https://github.com/sapientinc/HRM

TL;DR

WHAT was done? The paper introduces the Hierarchical Reasoning Model (HRM), a novel recurrent architecture inspired by the human brain. HRM features two interdependent modules operating at different timescales: a high-level module for slow, abstract planning, and a low-level module for rapid, detailed computation. This design enables significant computational depth while maintaining stability through a "hierarchical convergence" process. Training is made highly efficient by a one-step gradient approximation that bypasses Backpropagation Through Time (BPTT), and the model dynamically allocates resources using an Adaptive Computational Time (ACT) mechanism (see my series of posts on ACT).

WHY it matters? HRM fundamentally challenges the dominant "bigger is better" paradigm of large language models (LLMs). With only 27M parameters and trained from scratch on just ~1000 examples, it achieves near-perfect performance on complex reasoning tasks (like Sudoku-Extreme and Maze-Hard) where massive Chain-of-Thought (CoT) models completely fail. This suggests that sophisticated architecture and computational depth, rather than sheer scale, can be a more efficient and robust path to advanced AI reasoning. Furthermore, the model's learned internal structure spontaneously develops a brain-like dimensionality hierarchy, providing strong empirical validation for a brain-inspired approach.

Details

A New Blueprint for AI Reasoning

The pursuit of artificial general intelligence has long been intertwined with the challenge of complex reasoning. The prevailing approach, embodied by large language models (LLMs), has been to scale up model size and rely on techniques like Chain-of-Thought (CoT) prompting to externalize reasoning into sequential, human-readable steps. While powerful, this paradigm suffers from brittleness, high latency, and an insatiable appetite for data. A recent paper introduces the Hierarchical Reasoning Model (HRM), presenting a compelling alternative that suggests the key to reasoning may lie not in sheer scale, but in a more profound, brain-inspired architectural depth.

Keep reading with a 7-day free trial

Subscribe to ArXivIQ to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Grigory Sapunov
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share