LLM policy?

github.com

178 points by dropbox_miner 16 hours ago


SoftTalker - 15 hours ago

I think everything has a growing problem with LLM/AI generated content. Emails, blog posts, news articles, research papers, grant applications, business proposals, music, art, pretty much everything you can think of.

There’s already more human produced content in the world than anyone could ever hope to consume, we don’t need more from AI.

ddxv - 14 hours ago

I've seen an interesting politically motivated one. It didn't appear to be a bot, just a user from China:

https://github.com/umami-software/umami/pull/3678

The goal is "Taiwan" -> "Taiwan, Province of China" but via the premise of updating to UN ISO standards, which of course does not allow Taiwan.

The comment after was interesting with how reasonable it sounds: "This is the technical specification of the ISO 3166-1 international standard, just like we follow other ISO standards. As an open-source project, it follows international technical standards to ensure data interoperability and professionalism."

The politics of the intent of the PR was masked. Luckily, it was still a bit hamfisted. The PR incorrectly changed many things and the user stated their political intention in the original PR (the above is from a later comment).

a2128 - 12 hours ago

I think AI generated issues and pull requests may lead to the death of GitHub as the place for open-source projects. As Microsoft is hell-bent on shoving easy AI buttons in every interface including GitHub, and as more people try to pad their GitHub profile for their resume, the burden for open-source projects to filter the low-quality spam may grow immense and it may become worth to just switch to Codeberg or other platforms. Some LLM use will inevitably leak into these other platforms but the relatively high barrier (no built-in AI buttons), effective moderation banning obvious violators, and reduced usefulness for resume padding would make it negligible I think.

blutoot - 13 hours ago

Since the cat is out of the bag (i.e. there's no stopping of LLM-generated code / PR / issues now that vibe-coding is fairly universally accessible), the only thing in our (especially those who are custodians of software systems and products) control is to invest a lot more in testing and evals. Just plain opposing LLM-generated content as a policy is a losing battle at this point..

lofties - 15 hours ago

Sidenote, but I love that in a GitHub issue discussing banning the use of LLMs, the GitHub interface asks if there's anything I'd like to fix with CoPilot.

arpit15191 - an hour ago

I think there is soon going to be such implementations globally, if not within organizations or smaller groups. The ease of using v/s actual quality content is really the question should be answered when deciding LLM policy.

dropbox_miner - 16 hours ago

There seems to be a bot that routinely creates huge LLM generated Issues on runc github: https://github.com/containerd/containerd/issues/12496

And honestly, its becoming annoying

dSebastien - 13 hours ago

Prompt injection must be a really lucrative activity. I wonder how many hidden prompts will be detected on GitHub in the coming years and how many devs dare point LLMs at GitHub issues

zzo38computer - 14 hours ago

In stuff I maintain I hardly get anything at all, whether using LLMs or not, so it does not affect me much. Furthermore, I exclusively use the API, so the UI will not try to make you use Copilot and those things either (that is not the reason I use the API, although it is a side effect).

I do not use Copilot, Claude, etc, although I partially agree with one of the comments there, that using LLM for minor auto-completion is probably OK, as long as you can actually see that the completion is not incorrect (although that should apply to other uses of auto-completion too, even if LLM is not used; but it is even more important to check more carefully if LLM is used). I think it would be better to not accept any LLM generated stuff otherwise (although the author might use LLM to assist before submitting it if desired (I don't, but it might help some programmers), e.g. in case the LLM finds problems with it, that they will then have to review themself to check if it is correct, before correcting and submitting it; i.e. don't trust the results of the LLM).

It links to https://github.com/lxc/incus/commit/54c3f05ee438b962e8ac4592... (add .patch on the end of the URL if it is not displayed), and I think the policy described there is good. (However, for my own projects, nobody else can directly modify it anyways; they will have to make their own copy and modify that instead, and then I will review it by myself and can include the changes (possibly with differences from how they did it) or not.)

dropbox_miner - 16 hours ago

Curious to know if others are seeing a similar uptick in AI slop in issues or PRs for projects they are maintaining. If yes, how are you dealing with this?

Some of the software that I maintain is critical to container ecosystem and I'm an extremely paranoid developer who starts investigating any github issue within a few minutes of it opening. Now, some of these AI slop github issues have a way to "gaslight" me into thinking that some code paths are problematic when they actually are not. And lately AI slop in issues and PRs have been taking up a lot of my time.

xeeeeeeeeeeenu - 14 hours ago

Unfortunately, LLMs empower "contributors" who can't be bothered to put in any effort and who don't care about the negative impact of their actions on the maintainers.

The open-source community, generally speaking, is a high-trust society and I'm afraid that LLM abuse may turn it into a low-trust society. The end result will be worse than the status quo for everyone involved.

kijin - 13 hours ago

I have a simple rule: whenever I receive an issue or PR from an unfamiliar account, I skim the text and/or code and post 1-2 quick questions. If the submitter responds intelligently, then the issue warrants a more in-depth investigation. If not, the submitter clearly doesn't think that the issue is worth following up on, so it's not worth my time, either.

The more code a PR contains, the more thorough knowledge of it the submitter must demonstrate in order to be accepted as a serious contributor.

It doesn't matter if it was written by a human or an LLM. What matters is whether we can have a productive discussion about some potential problem. If an LLM passes the test, well it's a good doggy, no need to kick it out. I'd rather talk with an intelligent LLM than a clueless human who is trying to use github as a support forum.

alyxya - 11 hours ago

Maybe the solution is to counteract LLM pull requests with LLM first pass reviews. Obviously not a perfect solution, but LLMs should be good enough at detecting spam. This only needs to apply to first time contributors who haven't yet proven themselves.

jmyeet - 14 hours ago

Some years ago I read the Neal Stephenson book Anathem. SPOILERS: it has a version of the Internet called the Reticulum and one thing I remember is that it was filled with garbage. True information subtly changed multiple times until it was garbage. And there were agents to see through the garbage. I imagined this to be a neverending arms race.

Honestly, this is kind of where I see LLM generated content going where you'll have to pay for ChatGPT 9 to get information because all the other bots have vandalized all the primary sources.

What's really fascinating is you need GPUs for LLMs. And most LLM output is, well, garbage. What did you previously need GPUs for? Mining crypto and that is, at least in the case of Bitcoin, pointless work for the sake of pointless work ie garbage.

I can see a future in our lifetimes where a significant amount of our capital expenditure and energy consumption is used, quite simply, to produce garbage.

alexchantavy - 13 hours ago

As an open source maintainer, I don't have an issue with AI; I have an issue with low quality slop whether it comes from a machine or from a human.

The responsibility then is for an open source project to not be shy on calling out low quality/low effort work, have good integ tests and linters, and have guidance like AGENTS.md files that tell coding robots how to be successful in the repo.

vineyardmike - 15 hours ago

I've seen an uptick in LLM generated bug reports from coworkers. A employee of my company (but not someone I work with regularly) used one of the CLI LLMs to search through logs for errors, and then automatically cut (hundreds!) of bugs to (sometimes) the correct teams. Turns out it was the result of some manager's mandate to "try integrating AI into our workflow". The resulting email was probably the least professional communication I've ever sent, but the message was received.

The only solution I can see is a hard-no policy. If I think this bug is AI, either by content or by reputation, I close without any investigation. If you want it re-opened, you'll need to IRL prove its genuine in an educated, good-faith approach that involves independent efforts to debug.

> "If you put your name on AI slop once, I'll assume anything with your name on it is (ignorable) slop, so consider if that is professionally advantageous".

w10-1 - 11 hours ago

Policy is a decent filter for consensual candidates, but mostly ineffective against incentives and automation.

The issue is exported costs: whether submitters make reviewers work too hard for the contribution value.

The policy/practice should focus first on making reviewer/developer's work easier and better, and second on refining submitter skills to become developers. The same is true for Senior/Junior relations internally.

So the AI company that solves how to pare AI slop down to clean PR's would meet a real and growing need, and probably also help with senior/junior relations as well.

Then you could meet automation with automation, and the incentives are aligned around improving quality of code and of work experience. People would feel they're using AI instead of competing with it.

mannicken - 8 hours ago

Oh man this is all really hurting my brain. I wish I had a good response to this. I wish I had a magical simple response and a way to resolve this problem. On one end, I stop by curl's H1 profile regularly and watch people just drown them in garbage. It's like LLMs have unlocked a portal to a new kind of stupidity that was dormant before. Netsec always had a problem with people running nmap -sV -sC -sX -Pn -p1-65535 --max-rate -T1 scanme.nmap.org -oS out.txt and then just sending that out.txt with a subject "I FOUND A VULENARBILATEAY" And like, a script kiddie isn't really a person... everyone has a script kiddie inside of us. But it feels like we took this wonderful technology of neural networks and we did capitalism with it and unleashed it into a business of making it easier for people to be stupid but feel smart while doing it. I really like the idea that we might learn something about our own consciousness by simulating neural networks with computers but the more I see how it's done in reality the more I want to puke my brains out.

On the other hand, Matt Godbolt seems to use LLMs and I feel like I sure as hell wouldn't want to miss a PR from Matt fucking Godbolt. I mean even if I go full vanilla LLM-free I still am too addicted to using godbolt.org at this point and it was written partially with an LLM apparently.

Argh, maaan I don't know this is too fucking complicated of a problem for me to solve. Fuck, maybe let's just destroy all this technology and live as neofarmers raising chickens?

<Zero LLMs were used to write this post. In fact I went ahead and broke one GPU for every sentence I wrote just to make it harder for LLMs to compute.>

noduerme - 12 hours ago

>>> the entire issue description contains so much unneeded (and probably incorrect) information that it'd be better if they just provided their LLM prompt as an issue instead

When it's put this way, it seems a lot like the problem of people walking into doctors' offices with certainty that they know their own diagnosis after reading stuff on Reddit and WebMD.

What this post actually amounts to, indirectly, is a plea to trust human expertise in a particular domain instead of assuming that a layperson armed with random web pickings has the same chance as an expert at accurately diagnosing the problem. This wastes the expert's time and just increases mistrust.

The exceptions where Reddit solves something that a doctor failed to solve are what infuse the idea of lay online folk wisdom with merit, for people desperately looking for answers and cures. Makes it impossible to impose a blanket rule that we should trust experts, who are fallible as well.

The problem is societal. It's that if you erode trust in learned expertise long enough, you end up with a chaos of misinformation that makes it impossible to find a real answer.

A friend of mine who died of lung cancer recently, in his last days became convinced that he'd gotten it because of the covid vaccine (despite being a lifelong smoker, whose father had died of it at 41). And in every individual case you say, well, I don't want to disabuse someone of the fantasy they've landed on.

This is a devastatingly bad way to raise a generation, though. Short-circuiting one's own logic and handing it over to non-deterministic machines, or randos online... how do we expect this to end?

benatkin - 14 hours ago

The issue expresses doubt about a policy specific to LLMs being accepted. I think the way to go might be to accept that the bar for outside contributions should unfortunately be higher. It doesn't take an LLM to have a glut of low quality contributions, it just takes an incentive and some attention, as we've seen with Hacktoberfest: https://news.ycombinator.com/item?id=31628342

bsder - 14 hours ago

A much bigger problem is that when an AI/LLM coughs up code, you have absolutely no idea what the copyright or license is.

shashankg09 - 13 hours ago

This is a poor take. We need to stop slop and low effort issues/PRs. Stopping AI generated code is a lost battle because detecting that in high quality work is impossible.

denmark1 - 12 hours ago

[dead]

denmark1 - 13 hours ago

[dead]

ninetyninenine - 14 hours ago

This is a minor issue. The big issue comes when we start complaining about code that's not generated by AI. When hand coded stuff becomes buggier than AI stuff.

CGamesPlay - 14 hours ago

I make a lot of drive-by contributions, and I use AI coding tools. I submitted my first PR that is a cross between those two recently. It's somewhere between "vibe-coded" and "vibe-engineered", where I definitely read the resulting code, had the agent make multiple revisions, and deployed the result on my own infrastructure before submitting a PR. In the PR I clearly stated that it was done by a coding agent.

I can't imagine that any policy against LLM code would allow this sort of thing, but I also imagine that if I don't say "this was made by a coding agent", that no one would ever know. So, should I just stop contributing, or start lying?

[append] Getting a lot of hate for this, which I guess is a pretty clear answer. I guess the reason I'm not receiving the "fuck off" clearly is because when I see these threads of people complaining about AI content, it's really clearly low-quality crap that (for example) doesn't even compile, and wastes everyone's time.

I feel different from those cases because I did spend my time to resolve the issue for myself, did review the code, did test it, and do stand by what I'm putting under my name. Hmm.

NamlchakKhandro - 12 hours ago

Wouldn't you provide LLM friendly guidance material in your repo instead of spending effort on a pointless endeavor to ban them ?

brabel - 11 hours ago

It will be more interesting to see if and when AI becomes better than humans at coding, which seems to be coming close to reality, some projects may start accepting only AI contributions :D.