Nearly Right

Stack Overflow's collapse reveals the fragility of the programming knowledge commons

The platform that trained the AIs that replaced it was already dying from within

Monthly questions on Stack Overflow: 207,000 at peak. By late 2025: 3,700. The graph spread through programming circles in early January with the force of a death certificate, and the comments underneath it were brutal. Good riddance. Couldn't have happened to a meaner community. Let it die.

This hostility would have been unthinkable a decade ago. Stack Overflow was not merely useful; it was infrastructural, the site where you went when code broke at 2am and documentation failed you. Its orange logo became shorthand for how modern programming worked. You hit a wall, searched Google, clicked the Stack Overflow link, found your answer. The verb "stackoverflow" entered developer vocabulary alongside "google." When the founders sold for $1.8 billion in June 2021, the price reflected genuine centrality to the software industry.

Yet the graph tells a different story. The collapse did not begin with ChatGPT. The downward slope starts around 2017, five years before large language models became practical coding assistants. Something had already gone badly wrong.

The bureaucracy that ate itself

Jerry Pournelle's Iron Law of Bureaucracy holds that in any organisation, those devoted to the organisation itself will eventually take control from those devoted to its goals. Stack Overflow offers a textbook illustration.

The site launched in 2008 with a novel premise. Unlike the chaotic forums that preceded it, Stack Overflow would build a curated knowledge base. Questions would have definitive answers. Duplicates would be merged. Poor questions would be downvoted into oblivion. The founders, Joel Spolsky and Jeff Atwood, positioned this explicitly as the anti-forum: the answer to your problem mattered more than any conversation around it.

For years, this worked brilliantly. Strict moderation created signal amid noise. The gamification system attracted knowledgeable contributors who enjoyed both helping others and climbing leaderboards. By 2014, Stack Overflow had become the default destination for programming questions, its answers appearing at the top of nearly every technical search.

Then the mechanisms that built the knowledge base turned against the community that created it.

Moderation intensified. Questions were closed as duplicates even when the linked answers failed to address the specific problem. Comments demanding better formatting accumulated beneath earnest queries from newcomers. The easy questions had all been answered; later arrivals found the barriers to participation rising and the rewards diminishing. A developer with 25,000 reputation points described watching questions get closed before he could respond, killed by moderators enforcing what they saw as quality standards. He came from Usenet, where the goal was simply helping someone who needed it. Stack Overflow had different priorities. The person asking was less important than the normalisation of the dataset.

The site's defenders argued that critics simply misunderstood the mission. Stack Overflow was never meant to be a help desk. It was building a reference work, and reference works require editorial control. The policies were documented and debated. Users who felt insulted by question closures were mistaking curation for hostility.

This defence has merit. But a reference work that cannot attract new contributors eventually becomes a museum. The moderators optimised for content quality at the expense of contributor experience, protecting existing knowledge while choking off the generation of new knowledge. The bureaucracy had begun serving itself.

When the gatekeeper changed the locks

A former Stack Overflow employee offered a crucial detail in the discussion thread. For most of the site's history, the vast majority of visitors arrived via Google. Not search engines generally. Google specifically. This dependency was baked into the design from the start.

The symbiosis resembled a flower and its pollinator. Google directed millions of programmers to Stack Overflow's answers. Stack Overflow's existence improved Google's results for technical queries. Both benefited. The arrangement seemed permanent.

Then Google began changing how it surfaced results. The exact mechanisms remain opaque, but the effect was clear: Stack Overflow started appearing for unanswered or poorly answered questions rather than the canonical solutions that had made it valuable. Meanwhile, scrapers cloned Stack Overflow content onto ad-laden sites that Google often ranked higher than the original. The search engine that had served as Stack Overflow's primary interface was becoming unreliable.

Stack Overflow had outsourced its user interface to another company. When that company's priorities shifted, there was no fallback. The site had never developed strong internal search, or community features that might keep users engaged, or any relationship with users that did not depend on Google delivering them to the front door.

The pattern echoes ecological collapse. A keystone species changes its behaviour; the ecosystem it supported unravels with surprising speed. Stack Overflow had built its entire architecture around a single distribution channel. That channel's gradual deterioration left the community exposed to whatever disruption came next.

The machines trained on human frustration

The disruption arrived in November 2022. ChatGPT launched with fluency that stunned even researchers who had watched language models develop. Within months, programmers discovered these systems could answer many coding questions instantly, drawing on training data that included millions of Stack Overflow posts.

The irony cuts deep. Stack Overflow's content was licensed under Creative Commons, explicitly permitting reuse with attribution. The AI companies scraped this corpus and built tools that replaced the need for the original source. Contributors who had answered questions to help others had inadvertently helped train systems that would make their contributions unnecessary. The platform generated the training data for its own extinction.

The LLMs offered something Stack Overflow could not: patience. You could ask a stupid question without fear of downvotes. You could phrase your problem clumsily without someone demanding you show your work. You could iterate through failed attempts without anyone closing your thread. The machines inherited the knowledge that humans had painstakingly accumulated, then presented it without the social friction that had accumulated alongside.

The acceleration after ChatGPT's release was dramatic. Question volume, already declining, fell off a cliff. By early 2024, the site was receiving fewer monthly questions than in its first year of operation. The LLMs had not merely provided an alternative. They had revealed how much of Stack Overflow's remaining traffic had been reluctant usage by people who simply had nowhere else to go.

The generation that will not share

A contributor named 0xfaded offered a melancholy reflection. Years earlier, he had published an elegant algorithm for finding the closest distance between an ellipse and a point. He considered it his finest contribution to human knowledge. He posted it on Stack Overflow because that was where people who needed such answers would look.

The solution was cited in academic papers and incorporated into game engines. People reached out regularly to thank him or share their applications. Then the messages stopped.

He tested various LLMs with the same problem. None reproduced his method. They generated different approaches, often unstable ones his solution had been designed to avoid. The knowledge existed in Stack Overflow's archives, but the systems trained on that archive had not learned to surface it reliably.

Where would such gems be shared now? He did not know. Stack Overflow had been the obvious venue, the place programmers went when they had both questions and answers to offer. No equivalent exists today. The LLMs do not create community. They extract from existing knowledge without generating new collective understanding.

This points toward a troubling dynamic. Large language models train on human-generated content. But their existence reduces the incentive to generate that content publicly. Questions asked to ChatGPT leave no record. Solutions discovered through AI assistance are not shared on forums where others might find them. Knowledge moves from commons to black box, from searchable archive to ephemeral conversation.

The feedback loop has grim implications. If public technical discourse diminishes, future language models will have less material to train on. They will increasingly train on their own outputs, a process researchers have shown degrades quality over generations. The AI systems that killed Stack Overflow may eventually be limited by its absence.

What remains

Stack Overflow still exists. The archives remain searchable, a fossil record of problems programmers faced between 2008 and roughly 2023. The company has attempted various pivots: AI-assisted answering, enterprise tools, partnerships with the very language model companies that disrupted its core business. Whether any of these will sustain the organisation remains unclear.

The broader question is what takes its place as a generator of public technical knowledge. Discord servers and GitHub issue trackers absorb some demand, but these are fragmented and poorly indexed. Reddit maintains activity but lacks structure. No obvious successor has emerged, and the LLMs cannot fill the gap they created.

Perhaps the lesson is that knowledge commons are fragile constructions. Stack Overflow solved a coordination problem, giving developers a single place to both ask and answer questions. The site's strict moderation created costs that eventually outweighed benefits, but the alternative is not a better-moderated forum. The alternative is no forum at all—just each programmer alone with their AI assistant, solving problems no one else will ever learn from.

The graph approaches zero. The comments mix schadenfreude with unease. Programmers remember being insulted by moderators and celebrate their decline. They also remember finding answers that seemed like magic, solutions from strangers who took time to explain not just what worked but why. That second experience has no replacement.

What Stack Overflow built was imperfect, often unpleasant, ultimately unsustainable. What replaces it is more convenient but less generative. The machines learned from human collaboration, then made that collaboration unnecessary. Whether they can continue learning without it remains an open question—one that, unlike the problems Stack Overflow used to solve, has no accepted answer waiting to be found.

#software development