AI didn’t stackoverflow. Its own toxic environment did.
Learning to code, and run into an issue that is extremely common but you dont know that because again, youre learning to code? Answered before, removed
Have an issue that you dont know exactly how it came about, and others have been talking about it elsewhere because its a relatively new issue but is pervasive? Duplicate, removed
Have an issue because you aren’t God’s gift to coding? Verbally assaulted in the responses, removed.
It was a shit website dominated by the worst of computer science.
The strict moderation is the main strength of it. Makes it so much easier to find useful answers and for those answers to be refined.
Most of the toxicity I’ve seen surrounding the site has been from people upset that they were asked to improve their search skills, asked to improve their question to be more useful for those answering them, or simply demanding that people answer to their deadlines.
It’s hard to take complaints about the site seriously when so many of them seen to come from entitled arseholes being offended at being asked to not be arseholes.
Couldn’t they just merge posts if they were duplicates and avoid the whole toxic environment?
I’m sure they could have coded it like that.
It’s a mixture of so many common questions already having existing answers, the toxic community, and AI. There was already a decline happening before AI appeared, but it got much quicker after chatgpt released.
tbf, you have to compare it to what came before - old style forums dedicated to any given language, framework, or technology. Not only was stack better at consolidating questions and answers, it significantly cut down on the verbal abuse levelled at new coders. [Closed as Duplicate] might be offputting - but it is still better than 3 pages of insults and squabbling because you asked how to close vim.
[Closed as duplicate]
just kidding, but this was a huge issue.
You may notice that their decline starts before LLM’s rise begins. Cultivating an environment where specialists collect power through hostility to understandable mistakes and good faith questions was their real undoing, and it happened years ago.
Blaming AI is such a sad excuse.
By their own chart, StackOverflow’s decline started at least five years before the popularity of LLMs skyrocketed.
Meanwhile, as a developer I feel like I wasn’t using Stack Overflow at all by 2020.
yea that’s more my reading AND experience
“This question has already been answered by several programming textbooks. Why are you asking questions about it?”
Yeah, also [you posted bad question]
Something maybe being missed in this, is that stack overflow was also a community. It wasn’t just getting a question answered. If you are a specialist in some domain, you would be returning to the same community and get to know others asking question and others answering.
And that community was exploited, entirely.
I think a part of whats happening here has been a betrayal of trust when it comes to community building.
While the downturn went from gradual to rapid when ChatGPT released in late 2022, it looks like they’d been on a steady / gradually accelerating decline for a few years beforehand - they lost all their gains from 2012-onward before GenAI really became a thing.
Anyone know why? What was killing StackOverflow slowly before GenAI killed it quickly?
I think a big reason for StackOverflows downfall where the intransparent and often outright infuriating moderation decisions.
When people started answering questions like it was reddit. SO was already well in decline. AI just finished it off.
It looked pretty steady, normalized to a range of people with highs and lows, until the AI hit. SO was an early adopter of AI.
It’s clearly declining from 2018. They weren’t that early.
I believe software communities migrating to closed, volatile, private groups in discord or similars contributed to the issue
So people are not asking new questions. How will AI learn anything new then? 😅
So people are not asking new questions.
They are asking new questions. They’re just asking them directly to the AI interface.
The bigger problem is that they’re not getting new answers.
The vast majority of traffic to stack was always people searching for a solution to a problem, finding the question already asked, finding an accepted solution, using that solution, and leaving without contributing anything. (1)
A significantly smaller proportion of the traffic would be users who logged in, then voted on questions or answers they found useful. (2)
Smaller still, people who failed to do any real research prior, and asked a question that would inevitably be marked a duplicate. (3)
A tiny fraction would be people asking genuinely novel questions. (4)
And finally, the hardcore power users who would do all the actual work of sorting, moderating, and answering questions. (5)
We can imagine this traffic probably followed a distribution something like this:
x
x
x
x
x
x
x
x
x x
x x
x x
x x
x x x
x x x
x x x x
x x x x x
1 2 3 4 5Now that we have LLMs,
(1) LLMs are “better google”. This traffic will be significantly reduced, almost to nothing. That is what the OOP graph is really showing.
(2) This traffic will also be reduced almost to zero, though there will be some holdouts who keep showing up to vote.
(3) LLMs are pretty decent at figuring out how to rephrase things, so people asking obviously duplicate questions will, again, be greatly reduced. Every once in a while, someone’s LLM will fail to translate their question into the question someone else asked on stack, and they will show up and ask their duplicate question, feeding into the main usefulness of duplicate questions - identifying different ways people phrase the same question so that these can all point to the same solution. These new phrasings will be picked up by LLMs, reducing the need for similar rephrasings in the future.
(4) Many of the non-duplicate questions asked on Stack are basic things that someone new to a language, library, or framework might ask. “Hello world” level syntax questions which can be answered via rtfm. These can now easily be answered via LLM. The remaining questions will be about weird bugs specific to particular software versions, or how to integrate arcane parts of different frameworks, etc, which will continue to feed LLMs.
(5) This traffic will probably be largely unimpacted, and these users user experience will probably improve as they now have less junk to deal with. Fewer idiots making vaguely racist remarks in the comments. Fewer duplicate or rtfm questions. A higher signal to noise ratio of interesting novel puzzles that their particular brand of autism thrives on.
So the new graph of traffic post-LLM might look something like this:
x x x x x
1 2 3 4 5Admittedly, this might be hard on Stack’s business model of deriving ad revenue from web traffic… but as far as its actual purpose of providing answers to questions and providing a good experience to actual experts, it should continue to function just fine - if not better.
I think you’ve got it backwards, actually, a lot of power users publicly announced their departure when AI was introduced to the site and that probably lead the lower level users to leave as well since they could no longer get good answers.
Jon Skeet answered this question less than a week ago. Stack lives on.
Can you link to some power users publicly announcing their departure?
No offense, but that doesn’t sound like a good use of my time.
Fair but without data it seems questionable. It seems like a pretty big backpedal to not even half ass list off some names of people you remember. But I waste too much time here too
I’m a little confused, I went to SO and looked at how many questions were asked in the last 24 hours and counted ~250, which would track much higher than the 3k or so that the graph shows. Are these questions that don’t get deleted or removed? Or questions that get answered?
Idk, the source of the data is here:
https://data.stackexchange.com/stackoverflow/query/1926661#graph
Original automatically updating graph: https://data.stackexchange.com/stackoverflow/query/1926661#graph
ty
I think we’ll soon see exactly how good “reasoning” models actually are, as the next generation of problem-havers uncover things that were never discussed on SO or elsewhere in-depth.
Stack Overflow was the “most popular” Q&A site precisely because it was the site Google consistently elevated to the top of the search results queue. The thing that’s killing Stack overflow isn’t GenAI, it’s Google results deprioritization.
More often than not, I currently see Reddit at the top of my results queue in Google when I go fishing for online tech support answers.










