Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> Look at popular projects -- a few minutes after an issue is filed they have sometimes 10+ patches submitted. All generating PRs and forks and all the things.

I think this is a really important point that is getting overlooked in most conversations about GitHub's reliability lately.

GitHub was not designed or architected for a world where millions of AI coding agents can trivially generate huge volumes of commits and PRs. This alone is such a huge spike and change in user behavior that it wouldn't be unreasonable to expect even a very well-architected site to struggle with reliability. For GitHub, N 9s of availability pre-AI simply does not mean the same thing as N 9s of availability post-AI. Those are two completely different levels of difficulty, even when N is the same.



Not even talking about how useless it is to create tens of PRs to solve the same issue.

But GitHub karma botting is a thing now.

Remember those elitist ppl who removed answers on stackoverflow coz their answer is better with 90000 answers?

Yup, now they are on GitHub farming karma with bots.


yeah this is indeed a good insight. Back in the days, who would expect so many bots to "review" code and leave overly verbose comments under every PR in a popular repo?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: