The Paper That Couldn't Add: Neural Computers and the Problem They Cannot Solve
A preprint published this week argues that AI should stop using computers and become one instead. Meta AI and KAUST have built two working prototypes — NCCLIGen, which handles terminal interfaces, and NCGUIWorld, which navigates a desktop GUI — and the paper's core bet is that learned runtimes will eventually displace conventional instruction execution as the dominant computing substrate. That bet has a problem the authors named themselves: symbolic stability.
Symbolic stability is the ability of a system to produce consistent logical outputs across runs. Conventional computers have it because they execute explicit symbolic instructions. The prototypes built by Mingchen Zhuge and 19 co-authors — running on a Wan2.1 diffusion transformer, trained on roughly 1,100 hours of terminal recordings and 1,510 hours of desktop sessions — do not. They can render an interface with high fidelity. They cannot reliably perform two-digit arithmetic. Multi-step tasks cause behavioral drift. The authors are direct: current instantiations are exceptional renderers but fragile reasoners, as one independent analysis put it.
Here is the paradox the paper does not fully confront: symbolic instability is simultaneously the feature that makes this direction interesting and the reason it cannot yet be deployed anywhere that matters. The applications the roadmap envisions — medical devices, financial systems, real-time control — require exact logical consistency from the first deployment. A system that drifts on long task chains or produces different outputs from identical inputs cannot satisfy the requirements of any reliability-critical environment. The paper's three named open challenges — routine reuse, controlled updates, and symbolic stability — are not future upgrades. They are prerequisites for the applications the authors are ultimately pitching. Behavioral drift compounds the problem: the paper identifies composition without drift as a separate unsolved issue, meaning long chains of operations accumulate errors that cannot be bounded.
The research is honest about the distance. The long-term target is a Completely Neural Computer (CNC) — a system that is Turing complete, universally programmable, behavior-consistent, and semantically machine-native, meeting four strict criteria that distinguish it from conventional Von Neumann architecture. To get there, all three named challenges must be solved simultaneously, none have solutions in this paper, and the authors themselves estimate a mature CNC would require a substrate in the 10 trillion to 1,000 trillion parameter range — three to four orders of magnitude beyond current frontier models. Zhuge has described this target as "sparser, more addressable, and a little more circuit-like" than current dense MoE systems.
The author list includes Jürgen Schmidhuber, whose decades-long argument that learned machines should replace programmed ones gives the paper directional weight even when the specifics are provisional. The paper has a GitHub repository (data pipeline only, not the models) and a research blog on the project's site. The claims — 19 authors, April 7 submission date, prototype names, training data volumes, the three open challenges — are all confirmed against primary and independent sources. As one analysis notes, the gap between these prototypes and a deployable CNC is not a product cycle away — it is a fundamental research problem that this paper names clearly but does not solve.
For hardware and semiconductor teams, the interesting question is not whether this specific paper will ship but whether the trajectory it describes changes silicon priorities. A learned runtime that aims to make the model itself the running computer — maintaining full execution state as activations rather than instructions — would require different memory hierarchies than a CPU fetching from DRAM. Compute-to-memory ratios in future accelerators would shift. Update efficiency during learning — modifying a running neural computer's behavior without retraining from scratch — becomes a first-class constraint rather than a training-time concern. Current GPU architectures are not designed for that workload, and no one is building the silicon that would be.
What would kill this story: if the symbolic reasoning barrier is fundamental rather than a scaling problem. If the incompatibility between holistic numerical semantics and the local combinatorial semantics of conventional computation is not a challenge but a category, this stays an intellectual exercise. If behavioral drift cannot be bounded in any practical training regime, the gap between current prototypes and a usable CNC does not close in a product cycle.
What would make it matter faster: if any of the three open challenges demonstrates tractability at near-term compute. That demonstration does not exist in this paper. The paper argues the direction is worth taking seriously. Whether it survives contact with an actual deployment requirement is a different question — and the one the authors have not yet answered.
The source: arXiv:2604.06425, "Neural Computers," Mingchen Zhuge et al., submitted April 7, 2026. GitHub (data pipeline) and blog at metauto.ai.