▲ | topspin 4 days ago | |||||||||||||||||||||||||||||||||||||||||||||||||
> "After completing the T steps, the H-module incorporates the sub-computation’s outcome (the final state L) and performs its own update. This H update establishes a fresh context for the L-module, essentially “restarting” its computational path and initiating a new convergence phase toward a different local equilibrium." So they let the low-level RNN bottom out, evaluate the output in the high level module, and generate a new context for the low-level RNN. Rinse, repeat. The low-level RNNs are iterating backpropagation while the high-level is periodically kicking the low-level RNNs to get better outputs. Loops within loops. Composition. Another interesting part: > "Neuroscientific evidence shows that these cognitive modes share overlapping neural circuits, particularly within regions such as the prefrontal cortex and the default mode network. This indicates that the brain dynamically modulates the “runtime” of these circuits according to task complexity and potential rewards. > Inspired by the above mechanism, we incorporate an adaptive halting strategy into HRM that enables `thinking, fast and slow'" A scheduler that dynamically balances resources based on the necessary depth of reasoning and the available data. I love how this paper cites parallels with real brains throughout. I believe AGI will be solved as the primitives we're developing are composed to extreme complexity, utilizing many cooperating, competing, communicating, concurrent, specialized "modules." It is apparent to me that human brain must have this complexity, because it's the only feasible way evolution had to achieve cognition using slow, low power tissue. | ||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | username135 4 days ago | parent [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
As soon I read the hlm/llm split, it immediately reminded me of the human brain. | ||||||||||||||||||||||||||||||||||||||||||||||||||
|