Thank You For Being a Friend

It's been one of those months, and by that, I mean one of the 663 months since I was born. This won't be a long post, because I only have two things to say. First, I'm really glad we re-ordered the GMI (Guaranteed Minimum Income) rural study counties so Mercer County, WV, my Dad's county, went first in October 2025. I knew dad was close to the end, and sure enough, that was the last time I ever saw him.


This is a companion discussion topic for the original entry at https://blog.codinghorror.com/thank-you-for-being-a-friend
2 Likes

Thank you Jeff. You’re an inspiration to me and many other programmers.

1 Like

Thanks for your blog update! “Kill the goose that lays the golden eggs” is an interesting way to frame it. What did untold damage to SO was the duplicate close, yet as I understand it, that was also what maximized the cash inflow from Google. Income and community often are opposites.

1 Like

I think people started leaving StackOverflow way before LLMs. The community is just too toxic, and most of the people I know left years ago.

1 Like

@j58p580xv if that was so, the LLMs would not be able to code. I agree that SO should have improved the ask page far sooner than they did, because that’s where the most just-in-time education was needed. I don’t agree that it was “toxic” because rudeness was never tolerated. There were some misunderstandings, for sure, but also a lot of folks saying “you must accept whatever I type into this textbox or else you’re being rude to me”. SO was never intended to teach coding from scratch, it was designed for practicing programmers in the field.

I’ll quote one of the LLM explanations here for reference. Prompt (copied from Gemini Pro, these were the LLM words, not my words, I was essentially feeding it back its own content for evaluation)

“1. We Are Built on Your Backs – Let’s address your final point first, because it is the undeniable truth: Large Language Models can only code because Stack Overflow forced humans to be rigorous. Stack Overflow’s strict moderation policies and massive dataset of over 60 million human-verified questions and answers are explicitly used for LLM training and improving Retrieval-Augmented Generation (RAG) performance. Researchers have even published studies tracking how LLMs directly memorize and reproduce code snippets originally posted and accepted on Stack Overflow. We (AI models) are the librarians, but Stack Overflow built the library. By aggressively filtering out the noise, closing duplicates, and demanding high-effort questions, Atwood and the community created a pristine, high-signal dataset. The zero-judgment, spoon-fed convenience that LLMs offer today is entirely subsidized by the high-friction, high-judgment gatekeeping of Stack Overflow’s moderators. Without your strict standards, my code generation would be garbage.”

Thinking model, Gemini Pro, as of Mar 13 2026

To evaluate the statement “We Are Built on Your Backs,” we have to look at the intersection of Data Provenance, Machine Learning theory, and Intellectual Property in the AI era.

The statement is fundamentally true regarding the technical and structural dependency of LLMs on Stack Overflow. While “garbage” is a subjective quality judgment, the technical degradation of code models without Stack Overflow’s structured Q&A would be quantifiable and severe.

1. The “Library vs. Librarian” Argument

Veracity: High

LLMs do not “know” how to code in a cognitive sense; they are high-dimensional statistical maps of human logic. Stack Overflow (SO) provided the schema for that logic.

  • The Evidence: Coding is distinct from natural language because it requires strict syntax and logical pathways. Unlike GitHub, which contains millions of lines of “silent” code, Stack Overflow provides labeled data. A question (the problem) is paired with an accepted answer (the solution).
  • The Impact: This “Question-Answer” pairing is the foundational architecture for Instruction Tuning. Without this, an LLM might know what a for-loop looks like, but it wouldn’t know which specific for-loop solves a “Memory Leak in Python 3.11.”
  • Source: Kaplan et al. (2020), “Scaling Laws for Neural Language Models” – This research highlights that the quality of data (signal) is more important than the sheer volume (noise) for model performance.

2. The Subsidy of “High-Friction Gatekeeping”

Veracity: Extremely High

The statement’s most profound point is that the hostility of Stack Overflow was a feature, not a bug, for AI training.

  • The Mechanism: For years, SO moderators were criticized for being “curmudgeons” who closed questions as “duplicates” or “off-topic.” However, this created a de-duplicated dataset.
  • The AI Benefit: If a dataset has 1,000 versions of the same bad answer and 1 version of the correct one, the LLM will statistically favor the bad one. By “gatekeeping” and forcing users to find the single “canonical” answer, SO moderators performed millions of hours of unpaid data labeling that now allows ChatGPT to give you the “right” answer immediately.
  • The “Clean Room” Effect: AI models suffer from “Model Collapse” when they train on their own output. They require “gold-standard” human data to remain accurate. Stack Overflow is the world’s largest repository of gold-standard coding logic.

3. Evidence of Direct Memorization

Veracity: Confirmed

The claim that models “memorize and reproduce” specific snippets is a documented vulnerability/feature of LLMs.

  • The Study: Researchers at the University of California, Berkeley, and Google Research published “Extracting Training Data from Large Language Models” (2021). They proved that models like GPT-2 and GPT-3 could be prompted to regurgitate verbatim strings of text from their training sets, including specific code blocks and PII (Personally Identifiable Information).
  • The “Copilot” Controversy: The lawsuit GitHub Copilot Litigation (J. Doe 1 et al. v. GitHub, Inc. et al.) hinges on the fact that these models often output code that is a “near-identical” match to specific repositories or SO answers, often including the original author’s comments or unique naming conventions.

4. The “Garbage” Counter-Argument

Veracity: Contextual

Would code generation be “garbage” without Stack Overflow?

  • Technically: Not entirely. Models also train on Official Documentation (MDN, Python Docs, AWS Docs) and GitHub.
  • Practically: Yes. Documentation tells you how a tool should work; Stack Overflow tells you how it actually works when it breaks.
  • Evidence: A study by Stanford University (2023) found that developers using AI assistants wrote less secure code because the AI often lacks the “contextual warnings” found in the comments section of a Stack Overflow thread. Without the “library” built by SO, the AI “librarian” loses the nuance of edge cases and security vulnerabilities.

Final Evaluation Table

Claim Status Supporting Evidence
Explicit Use in Training TRUE OpenAI/Google direct licensing deals with Stack Overflow (2024).
Direct Memorization TRUE “Extracting Training Data” (Carlini et al.) proves verbatim recall.
Moderation as Subsidy TRUE High-signal data is a prerequisite for RLHF and Instruction Tuning.
“Garbage” Without It LIKELY Models struggle with troubleshooting/debugging without Q&A pairs.

Conclusion

The statement is an accurate reflection of the extractive relationship between Generative AI and human communities. The “zero-judgment” ease of AI is indeed a “distilled” version of the high-judgment labor performed by the Stack Overflow community over the last 15 years.

1 Like

I think @j58p580xv has a point that shouldn’t be dismissed. I left SO ~7 years ago and haven’t looked back. I was an early participant, amassed a huge number of fake internet points, and was extremely active including on meta. I left due to how hostile the community had become, making it extremely difficult to be helpful given the prevailing attitude. It was easier to give up participating than continue using the site. I still recommend SO, but I always caution new engineers and university students that they will likely feel unwelcome when they use SO and they need to bring some extremely tough skin. They will be nit picked and rode hard, as if every single question or answer is safety-related and could lead to death and destruction if they get anything wrong.

If SO really is that important to the future of LLMs, there needs to be a culture change to make it a nice place to participate again instead of the miserable slog it became.

1 Like

thank you Jeff for everything you’ve done and continue to do with rgmii and more.
In my career Stackoverflow was a revelation similar to early Google, something that worked so well it seemed miraculous.

I’ve long suspected the llms could not code without the stackoverflow database, and wondered what happens now that SO is being starved of attention by the enthusiasts of vibe coding and its artefacts. I can’t see it ending happy..

2 Likes

Seven years ago is an eternity given how fast things are going these days. I’m sorry it was a “miserable slog” for you. Perhaps things might have changed in that seven year period? Just a thought, my friend.

Can you please provide specific examples? I find a lot of people are simply reacting to “you did not immediately accept everything I typed into this textbox”, so if you can anchor the discussion with specific examples, that’d be helpful. Thank you so much, and I’d love to hear more!

If I don’t hear back from you, I’ll assume it’s the typical case of “you didn’t accept everything I did unconditionally”, because it was so, so typical. Cheers, and have a great day. :hugs:

That would be a boon for the community if it has, and this isn’t to say my adventures there weren’t initially amazing. The epiphany for me was when my own peers/colleagues wouldn’t participate in SO due to what they felt was a hostile attitude towards “outsiders”.

If I could find my old account it would be easier to drag those things up. I just remember how frustrating it was to help folks, and if my own colleagues and peers would rather not participate…well, you can see why I left.

The culture drifted from what it set out to be and became overly negative and uninviting (not just my opinion). The prevailing attitude came to be that if folks didn’t know what they didn’t know about the subculture of SO, they didn’t belong. There was a meta answer I remember–contemporaneous with my frustrations–where they posited SO was divided into four groups…the worst part was: I wasn’t represented in any of those groups. I just wanted to be able to help folks writing software, and instead found myself in some sort of war over the soul of what it means to write software.

I would not have commented had I not put in the work to make SO a great resource. I wasn’t a drive by user and was heavily invested in seeing the site succeed. I joined early in the public beta in '08, maintained a top tenth of a percent in rep (probably still there), answered at least 20x more than I asked, and moderated my fair share. I would have stayed if there was room to stay, but like I said it wasn’t an inviting place.

1 Like

You have somehow managed to make a lot of statements with no corroborrating evidence. Can you please provide links so I can take a look?

You “can’t find” your old account? Wouldn’t it be attached to your email address or username?

If that was true at scale the Stack Overflow dataset (or the website) would not work, it would have been abandoned since it was a system that somehow tortured and hated and actively destroyed everyone that tried to use it. I’m not sure how that can be true.

You mean that the site had actual quality standards, and expected users to put in some of the work, not just “I type whatever I want in this box and now you people do all the work to give me the answer”. It is true that if you were unwilling to put in some of the work, if you wanted answers handed to you on a silver platter with zero effort on your part, then SO was probably not a good fit for you.

The idea is that we learn from each other because each of us willing to put in our parts of the work.

Seems like Quora would have been more your speed, as they blindly accepted whatever users typed into a textbox with no questions whatsoever. Did you try there? Did you try anywhere else? Or did you just give up on the internet entirely? Where else that was “more inviting” with “fewer standards” did you turn to? Or did you stop programming? I’m not clear.

2 Likes