Hierarchical Reasoning Model (arxiv.org)
from yogthos@lemmy.ml to technology@lemmy.ml on 21 Jul 19:03
https://lemmy.ml/post/33479798

github.com/sapientinc/HRM

Hierarchical Reasoning Model is a new architecture that’s inspired by neural computation principles observed in the brain, such as hierarchical processing, temporal separation of neural rhythms, and recurrent connectivity.

The bio-inspired design demonstrates significantly improved efficiency and accuracy on complex reasoning tasks compared with current LLMs.

The HRM architecture is designed to achieve significant computational depth while maintaining stability and efficiency during training. It consists of two interdependent recurrent modules operating at different speeds.

The High-Level module operates slowly and is responsible for abstract planning and deliberate reasoning. The Low-Level module functions rapidly, handling detailed computations.

A dual-module system allows the HRM to perform sequential reasoning tasks in a single forward pass without needing explicit supervision of intermediate steps. The model is also designed to be Turing-complete, meaning it can theoretically simulate any Turing machine, overcoming the computational limits of standard Transformer models.

Another interesting feature is the use of one-step gradient approximation, which improves efficiency by avoiding the computationally intensive backpropagation through time method typically used for recurrent networks. Avoiding backpropagation offers a constant memory footprint, making the model more scalable.

The model also incorporates an Adaptive Computation Time mechanism, inspired by the brain’s ability to switch between fast, automatic thinking and slow, deliberate reasoning. The HRM is thus able to dynamically allocate computational resources based on the complexity of the task.

Despite having only 27 million parameters, the HRM achieves nearly perfect performance on difficult tasks like complex Sudoku puzzles and finding optimal paths in large mazes, areas where even advanced models using Chain-of-Thought (CoT) methods fail completely.

The HRM also outperforms much larger models on the Abstraction and Reasoning Corpus benchmark for artificial general intelligence. It achieved a 40.3% accuracy, surpassing models like 03-mini-high (34.5%) and Claude 3.7 8K (21.2%).

The model’s design means that its training phase is much cheaper as well. It can be trained effectively with a small number of examples (around 1,000) and does not require pre-training or CoT data.

HRM conducts computations within its internal hidden state space which is more efficient than CoT where reasoning is externalized into token-level language. The externalization process can be brittle and requires extensive data to work.

#technology

threaded - newest