← All Posts

AYA and the Dupoux-LeCun-Malik Paper — Part 10

18 Months Ahead of the Paper (And What That Means)

April 9, 20265 min readAyanami Hobbes

Let me be clear about what I'm claiming and what I'm not, because the temptation after reading a paper that validates your approach is to overstate the case, and I'd rather be credible than impressive. I'm not claiming we invented autonomous learning theory. Dupoux, LeCun, and Malik have been thinking about this stuff longer than we've been a company. They have decades of cognitive science and AI research behind their framework. We have 18 months of building things that kept breaking until we figured out why. I'm not claiming we implemented everything in the paper. We didn't. Learning from imagination, rich developmental scaffolding, meta-evolution—that's future work for us. Some of it is years away. What I am claiming is narrower: we independently arrived at the same architectural principles, from a different starting point, and implemented the core systems before the paper existed. That's either convergent evolution or a very specific coincidence. I'm betting on convergent evolution.

When multiple independent teams arrive at similar conclusions, it's usually because the problem forces the solution. The problem: AI systems that freeze after training fail in dynamic environments. They work great in demos. They work fine for the first few months of deployment. Then something changes—a payer updates requirements, a workflow shifts, a new edge case emerges—and the frozen system keeps doing what it was trained to do, which is no longer what it should do. You either retrain (expensive, slow, requires human ML teams) or watch performance degrade. The solution space is constrained: some combination of continuous learning from observation, continuous learning from action, and meta-control that coordinates the two. We discovered this by trying to deploy frozen systems into healthcare and watching them fail expensively. The paper discovers it through theoretical analysis of how biological organisms learn across evolutionary and developmental timescales. Different paths. Same destination. Same reason: the problem is real and the solution space is constrained.

The timeline, for those keeping score: September 2024, we start building. The initial design includes pattern-based learning and outcome feedback. Not because we read cognitive science papers—because frozen systems weren't working for healthcare workflows and we needed something that would. January 2025, we implement distributed meta-control. Multiple services handling confidence assessment, strategy selection, error recovery. The single System M approach felt too fragile for production, too many eggs in one basket. May 2025, pattern composition goes live. Level-based pattern hierarchy. Simple patterns scaffold complex ones. We call it pattern leveling, not Evo/Devo, but the concept is identical. October 2025, confidence calibration matures. Adaptive thresholds. Multi-factor scoring. The system learns when to act autonomously versus when to consult. March 2026, the Dupoux-LeCun-Malik paper drops. We read it. We recognize our architecture in their framework. We're slightly unnerved by how specific the alignment is. This series gets written.

What this means for us: credibility, roadmap validation, and competitive positioning. Credibility: when we talk to investors, partners, customers about our architecture, we can point to serious academic work that validates the approach. "We built something that aligns with where LeCun thinks AI needs to go" is a sentence that matters in certain rooms. We didn't get the memo beforehand—we arrived at the same place independently—but the validation is real. Roadmap validation: the paper identifies capabilities we don't have yet, specifically imagination-based learning, richer scaffolding, meta-evolution. Now we know these aren't just nice-to-haves. They're theoretically grounded improvements with serious people arguing they're essential for the long term. The roadmap has backing beyond our own intuitions. Competitive positioning: most AI companies are shipping frozen models with aspirational "continuous learning" claims in their marketing. We can say "here's the academic framework for what autonomous learning actually requires, here's how we implement the core systems, ask your other vendors the same questions and see what they say." That's a differentiated position.

What this means for the market, beyond just us: the paper isn't about validating our specific company. It's about where the entire field needs to go. Every AI company faces the same problem. Frozen models degrade in production. The solutions are limited: continuous retraining pipelines (expensive, slow, human-dependent, scales badly), pretend the problem doesn't exist (dangerous, eventually catastrophic), or build autonomous learning systems (hard to build, but scalable once you have them). The paper makes a strong case for option three. It provides theoretical grounding. It identifies the components: observation, action, meta-control, developmental scaffolding. Companies that build this will outcompete companies that don't. Not immediately—frozen models can be very capable at launch, and most customers don't dig into the architecture during sales cycles. But over time, as deployment conditions shift, as data distributions change, as new cases emerge that weren't in training data. We bet on autonomous learning 18 months ago. The paper suggests the bet was correct. The market will validate further.

What comes next: short term (2026), continue refining what we have. Better outcome classification. Richer pattern composition. Tighter integration between observation and action systems. There's still optimization to do on the core architecture. Medium term (2027), tackle the gaps the paper identifies. Learning from imagination requires a simulation layer—that's substantial engineering. Active curiosity requires careful design in regulated contexts where you can't manufacture cases for learning purposes. Long term (2028+), meta-evolution. Evolutionary optimization of the learning architecture itself. The paper says this requires millions of simulated life cycles. That's not happening overnight. But the architectural hooks can be laid now.

The honest claim after reading this paper: we identified the core problem independently, we arrived at similar architectural principles through different paths, we implemented the core systems before the paper existed, we have significant work remaining that the paper maps clearly, and the convergence suggests we're on the right path even if we're not at the destination yet. I'm not claiming genius. I'm claiming pattern recognition and willingness to build something different when the conventional approach kept failing. The serious thinkers validated the direction. That's all I've got.

Catch you bitches on the flip side.


Series Index

  1. When Yann LeCun Validates Your Architecture (And You Didn't Even Ask)
  2. System M Isn't a Single Brain (And Neither Is Ours)
  3. Outsourced Learning Is a Dead End (And We Said So First)
  4. The Confidence Calibration Problem Nobody Talks About
  5. Learning from Execution: The Feedback Loop Nobody Ships
  6. Pattern Composition: Simple Before Complex (Always)
  7. The Transparency Imperative: Why Logic Logs Aren't Optional
  8. Observation + Action: The Integration Nobody Gets Right
  9. Where the Paper Says We Should Improve (And They're Right)
  10. 18 Months Ahead of the Paper (And What That Means)

Referenced Paper

Dupoux, E., LeCun, Y., & Malik, J. (2026). “Why AI systems don’t learn and what to do about it: Lessons on autonomous learning from cognitive science.” arXiv:2603.15381

Licensed under CC BY 4.0. TRIZZ AI is not affiliated with the authors. All opinions are our own.

← Part 9: Where the Paper Says We Should Improve (And They're Right)