Why AI systems don't learn – On autonomous learning from cognitive science
aanet
73 points
22 comments
March 17, 2026
Related Discussions
Found 5 related stories in 50.5ms across 3,471 title embeddings via pgvector HNSW
- Why No AI Games? pavel_lishin · 67 pts · March 03, 2026 · 57% similar
- If AI has a bright future, why does AI think it doesn't? JCW2001 · 15 pts · March 06, 2026 · 57% similar
- Adults Lose Skills to AI. Children Never Build Them ndr42 · 90 pts · March 28, 2026 · 55% similar
- Stanford study reveals AI vision models invent images they never see LionTurtle13 · 36 pts · March 30, 2026 · 54% similar
- "Cognitive surrender" leads AI users to abandon logical thinking, research finds Bender · 68 pts · April 03, 2026 · 53% similar
Discussion Highlights (6 comments)
aanet
by Emmanuel Dupoux, Yann LeCun, Jitendra Malik "he proposed framework integrates learning from observation (System A) and learning from active behavior (System B) while flexibly switching between these learning modes as a function of internally generated meta-control signals (System M). We discuss how this could be built by taking inspiration on how organisms adapt to real-world, dynamic environments across evolutionary and developmental timescales. "
beernet
The paper's critique of the 'data wall' and language-centrism is spot on. We’ve been treating AI training like an assembly line where the machine is passive, and then we wonder why it fails in non-stationary environments. It’s the ultimate 'padded room' architecture: the model is isolated from reality and relies on human-curated data to even function. The proposed System M (Meta-control) is a nice theoretical fix, but the implementation is where the wheels usually come off. Integrating observation (A) and action (B) sounds great until the agent starts hallucinating its own feedback loops. Unless we can move away from this 'outsourced learning' where humans have to fix every domain mismatch, we're just building increasingly expensive parrots. I’m skeptical if 'bilevel optimization' is enough to bridge that gap or if we’re just adding another layer of complexity to a fundamentally limited transformer architecture.
jdkee
LeCun has been talking about his JEPA models for awhile. https://ai.meta.com/blog/yann-lecun-ai-model-i-jepa/
zhangchen
Has anyone tried implementing something like System M's meta-control switching in practice? Curious how you'd handle the reward signal for deciding when to switch between observation and active exploration without it collapsing into one mode.
tranchms
We are rediscovering Cybernetics
Frannky
Can I run it?