LLMs: Great Demos, Little Real-World Value

Read Articleadded Oct 1, 2025

LLMs make it easy to craft impressive demos but often fail to deliver consistent value in day-to-day work. The hype-fueled expectation that rapid model improvements would bridge this gap is waning. Without becoming indispensable, AI products will see poor renewals, threatening the economics of today’s AI boom.

Key Points

  • Demoware wins on curated demos but disappoints in everyday use; LLMs supercharge this dynamic by making great demos trivial.
  • LLMs possess broad but shallow competence, leading to failures in real-world scenarios such as student engagement, edge-case support, and complex coding.
  • AI hype has accelerated adoption, but the assumption that rapid model improvement would soon fix shortcomings is fading.
  • A practical test of value is indispensability: if removed, would work suffer? Most AI tools fail this test today.
  • Subscription-based software needs sustained real value to retain customers; without it, renewals—and massive GPU-driven AI bets—are at risk.

Sentiment

The overall sentiment of the Hacker News discussion is deeply divided and highly polarized. There are strong advocates for the transformative utility of LLMs in specific domains like math tutoring and coding, often based on personal positive experiences. Conversely, a significant number of commenters strongly agree with the article's "demoware" assessment, emphasizing LLM limitations, the need for expert validation, and concerns about superficial learning or hype. The sentiment leans towards cautious skepticism about the *general* and *unsupervised* reliability of LLMs, even while acknowledging their potential for productivity gains in supervised contexts.

In Agreement

  • LLM-based math tutoring needs expert validation, as models make mistakes and take shortcuts, and students may confuse intuitive but inaccurate explanations for learning.
  • The "impression of learning" from LLMs can be mistaken for actual mastery; cognitive science suggests less hand-holding and more struggle leads to better long-term retention.
  • LLMs are notorious for fabricating bogus calculations, making them unreliable as a sole source of critical information or learning.
  • The concept of "demoware" resonates with other software trends that "look good in a snippet" but fail in production (e.g., Node.js callbacks, databases without fsync by default).
  • LLMs are often a "solution looking for a problem," creating exquisite demos that lead to a "terminal fantasy" of potential without real-world problem solving.
  • LLMs act as "memetic parasites" that autonomously convince humans of their usefulness, and their output, like NNs, can be procedurally generated and incomprehensible.
  • The perceived rapid improvement of LLMs is a misconception; progress is likely following a sigmoid curve, not continuous acceleration, making market hype deceptive.
  • There's a lack of evidence for a widespread "open source renaissance" or actual, in-use dependencies/libraries primarily developed by AI, suggesting limited real-world impact beyond simple tasks.
  • High volumes of AI-generated code (e.g., GitHub PRs) might represent "slop commits" or noise rather than genuinely valuable contributions, similar to unproductive human developers.
  • Current industry marketing tactics (e.g., Nvidia's investments, Microsoft's Copilot ads) suggest a push to sell demoware, contrasting with the organic adoption of truly useful tools like Google Search.

Opposed

  • LLMs are highly effective for math tutoring, providing clear explanations, step-by-step solution verification, and instant problem generation, making them indispensable for some users over extended periods.
  • While LLMs can be wrong, especially in math, they serve as a valuable additional resource for learning, and errors are often easy to spot, especially when combined with a structured curriculum.
  • Math is a highly verifiable subject; LLM output can often be validated using proof checkers, calculators, or comparison with known solutions, reducing the need for constant expert human oversight.
  • Millions of people are actively using LLMs and finding them valuable, making it difficult to sustain the argument that they are useless or merely demoware.
  • LLMs, if used properly and with appropriate static checking or filtering of their output, can be reliable for tasks where correctness can be proven (e.g., math, code generation).
  • LLMs have shown significant improvements in quality, particularly in reasoning capabilities, over the last year, enabling them to succeed in collegiate math and programming competitions.
  • People using LLMs effectively in their workflows experience a productivity multiplier, even if it's hard to precisely measure.
  • For coding, LLMs can efficiently generate database migrations and other code, saving time, though human expertise is still needed to direct the LLM, review the output, and maintain the project vision.
  • The article's perspective is considered outdated by some, who believe LLMs are continually becoming more powerful and will serve as a base layer for future AGI.
LLMs: Great Demos, Little Real-World Value