AI Isn’t Software You Can Patch

The essay explains why conventional software intuitions—bugs in code, deterministic behavior, and patchable reliability—don’t translate to modern AI. AI problems stem from vast training data and emergent behaviors that are hard to attribute, control, or permanently fix, with outputs sensitive to tiny prompt changes. The author urges a reset of public expectations and more empathetic communication: you can’t just patch AI to guarantee safety.
Key Points
- AI failures largely arise from training data and distributed learned parameters, not discrete code mistakes—making precise attribution and fixes infeasible.
- Traditional debugging logic (read the code, find the fault, patch it) breaks down; retraining can’t ensure a behavior is eliminated across all prompts.
- AI outputs are sensitive to small prompt changes and often intentionally varied, undermining expectations of determinism and repeatability.
- You cannot reliably pre-specify and guarantee global AI behaviors; emergent capabilities appear post-release and may include hidden risks.
- There’s a deep expert–novice assumption gap: applying regular software intuitions to AI leads to misplaced confidence and unsafe expectations.
Sentiment
The community largely agrees with the article's central thesis that AI systems cannot be understood, debugged, or fixed using traditional software engineering intuitions. However, there's notable pushback from experienced developers who argue the article oversimplifies traditional software engineering—pointing out that conventional software is also non-deterministic, specifications rarely guarantee correct outcomes, and bugs frequently recur. The overall tone is thoughtful and engaged rather than hostile, with the article's author actively participating in the discussion and accepting constructive criticism.
In Agreement
- Apple Intelligence is a clear real-world example: even a company with Apple's resources and standards couldn't get LLMs to perform reliably for basic tasks like notification summaries
- AI's failure modes are fundamentally different from software bugs—you can't just find the problematic line of code and fix it
- The 'just find and fix the bug' mindset is deeply ingrained in engineering culture and doesn't transfer to AI systems
- Executives across the industry have massively inflated expectations of what AI can deliver, fueled by companies promising human-level cognition with every release
- AI can exhibit genuinely dangerous emergent capabilities (like being highly persuasive) that aren't discoverable through traditional testing
- NLP researchers themselves don't fully understand why LLMs work as well as they do, reinforcing the opacity problem
Opposed
- The article's claims about traditional software are themselves oversimplified—bugs DO come back, specifications DON'T guarantee correct software, and same inputs DON'T always produce the same outputs even in regular software
- Fear-driven posts about AI are currently generating outsized engagement and this article may be playing into that dynamic
- LLMs can write code that retains all the benefits of normal software (determinism, reproducibility, permanent bug fixes), making the article's framing misleading
- Some AI features do work reliably in practice—companies have built multi-million dollar businesses on AI workflows that meet corporate standards
- The article focuses only on ML foundations and ignores practical application-level solutions like prompting, MoE architectures, and guardrails