Discussion about this post

User's avatar
Liz Gee's avatar
8hEdited

Thank you for this excellent article. I agree. LLMs can generate language but thought is computational and requires discrete steps.

Expand full comment
Neural Foundry's avatar

The compression versus procedural distinction you're drawing here cuts deeper than most people realize. What makes reasoning architectural rather than parametric is exactly what you're describing: the need for persistent state, conditional branching, and explicit backtracking. Training teaches pattern completion, but reasoning requires control flow that operates outside of token-by-token generation. The evolutionary valley analogy is especially apt—optimization pressure naturally selects against carrying forward the very uncertainty that enables genuine exploration. When you collapse the search tree into a single path during training, you're essentially teaching the model to approximate outcomes without learning the search algorithm. It's the difference between memorizing chess positions and understanding how to evaluate them dynamically. This is why infrastructure-based reasoning isn't just more efficient, it's categorically different.

Expand full comment
4 more comments...

No posts

Ready for more?