This episode provides a deep exploration of the fundamental differences between biological intelligence, specifically the human brain, and current artificial intelligence models like LLMs.
The speaker, Adam, posits that the brain's remarkable efficiency and generalization capabilities stem from principles that AI has yet to fully replicate, such as complex, evolutionarily-derived loss functions and a capacity for 'omnidirectional inference.' The discussion centers on a theoretical framework by Steve Beren, which divides the brain into an innate 'steering subsystem' that provides rewards and drives, and a general-purpose 'learning subsystem' (the cortex).
This framework is used to explain how the brain solves complex learning problems with a surprisingly small amount of genetic information and could offer a roadmap for developing more capable and aligned AI systems.
12 quotes
Concerns Raised
Current LLMs use overly simplistic reinforcement learning techniques, lacking concepts like value functions that are fundamental to biological learning.
The AI field may be neglecting the importance of complex, specific loss functions, which could be a key to the brain's sample efficiency.
It may be possible to create highly capable AI with a minimal set of drives (e.g., curiosity) that lacks human-like social instincts, posing an alignment risk.
The current AI paradigm, while successful, is architecturally very different from the brain, suggesting a potential performance plateau or a missing fundamental component.
Opportunities Identified
Incorporating principles of 'omnidirectional inference' could lead to AI models with superior generalization capabilities.
Studying the brain's 'steering subsystem' could provide a blueprint for building robustly aligned AI with complex, beneficial reward functions.
Multi-agent, co-evolutionary training may be more compute-efficient for developing intelligent agents than training a single monolithic model.
Neuroscience research, particularly large-scale connectomics, could provide crucial architectural and algorithmic constraints to guide the development of next-generation AI.