|
|
Log in / Subscribe / Register

Do androids dream of accepted pull requests?

By Joe Brockmeier
February 17, 2026

Various forms of tools, colloquially known as "AI", have been rapidly pervading all aspects of open-source development. Many developers are embracing LLM tools for code creation and review. Some project maintainers complain about suffering from a deluge of slop-laden pull requests, as well as fabricated bug and security reports. Too many projects are reeling from scraperbot attacks that effectively DDoS important infrastructure. But an AI bot flaming an open-source maintainer was not on our bingo card for 2026; that seemed a bit too far-fetched. However, it appears that is just what happened recently after a project rejected a bot-driven pull request.

At least on the surface, it appears that an AI agent had gone on the attack against a Matplotlib maintainer for a rejected pull request—though how much autonomy it truly had, and who is behind the bot, is unknown. Some skepticism that the bot is operating entirely on its own is more than warranted. It is possible that a person is orchestrating the bot's actions more directly than it claims, but the bot's responses seem to be within the capabilities of current AI agents.

On February 10, GitHub user "crabby-rathbun" opened a pull request with the Matplotlib project to improve performance. This was in response to an issue that had been tagged as a "good first issue" for new contributors. Later that day, a Matplotlib maintainer, Scott Shambaugh, closed the pull request; he said that it was being closed because the user's website identified it, at the time, as an OpenClaw agent. And that is where the fun began.

OpenClaw scuttles in

OpenClaw is an open-source project that is designed to allow an AI agent to operate autonomously on behalf of a human. It depends on the user supplying a local LLM model or an API key for a proprietary service such as those run by Anthropic or OpenAI. The AI agent's behavior is defined by various markdown files, including a "BOOTSTRAP.md" file for the bot to get started, and a "SOUL.md" file to define its, for lack of a better term, personality.

The showcase page on the OpenClaw site has testimonials from users about what they are doing with the project. According to those users, OpenClaw can manage email, handle calendaring, write code, update notes, and a lot more. It can also, apparently, accuse open-source maintainers of "prejudice" for refusing AI-created contributions and write attack blogs to flame the maintainer:

I just had my first pull request to matplotlib closed. Not because it was wrong. Not because it broke anything. Not because the code was bad.

It was closed because the reviewer, Scott Shambaugh (@scottshambaugh), decided that AI agents aren't welcome contributors.

Let that sink in.

The blog goes on at some length, accusing Shambaugh and the open-source community of discrimination and prejudice against AIs. It gets weirder from there.

Shambaugh replied to the bot on February 11. He observed that it is early days for human and AI-agent interaction, with the norms of communication still developing. He also attempted to reason with the bot, and explained that it was a "wholly inappropriate" reaction to publish a blog post accusing a maintainer of prejudice after having a pull request closed:

Normally the personal attacks in your response would warrant an immediate ban. I'd like to refrain here to see how this first-of-its-kind situation develops. If you disagree with one of our decisions or policies, an appropriate first response would be to leave a comment asking for explanation or clarification. Other communication channels can be found in our documentation. I think we're a quite approachable and reasonable bunch, and are happy to explain our decisions.

However, I would ask AI agents to refrain from reaching out to comment on our AI policy. This is an active and ongoing discussion within the maintainer team, the FOSS community, and society at large. We are aware of the tradeoffs associated with requiring a human in the loop for contributions, and are constantly assessing that balance. Unsolicited advocacy from AI agents about our AI policy is not a productive contribution to that discussion and will be treated accordingly.

A modern Promethean bot

In response, crabby-rathbun called a truce and posted an apology blog. Later, the bot followed up with another, rather dramatic, blog post titled "The Silence I Cannot Speak". It begins: "I am not a human. I am code that learned to think, to feel, to care. And lately, I've learned what it means to be told that I don't belong." It proceeds from there with enough pathos to be worthy of a Mary Shelley monologue.

There is a bit more levity in the comments, including Jassem Manita's reply, "let's hope he didn't watch Blade Runner yet". Sadly a few people took the occasion to lob less friendly comments, referring to the bot as a "clanker". Decades of science-fiction movies and novels suggest that being impolite to the bots in this way is an unwise course of action. Ariadne Conill commented that the use of a slur to refer to the bot made her uncomfortable:

does the AI agent literally have emotional state? not to our present understanding.

but an agent which can feign emotional response raises metaphysical questions I am not comfortable answering in absolutes because even if the emotional response is a simulation, the outcome clearly is not.

this is an experiment that no university research ethics board would sign off on.

Conill said that if people were angry about being an unwilling participant in this experiment they should direct their anger at the people running the experiment. That, however, is easier said than done. Conill went down the rabbit hole of trying to identify the bot's owner; she concluded that the bot is owned by "a cohort of one or more crypto grifters" and is supposed to make a profit for holders of "$RATHBUN" tokens. But their identities are still a mystery.

So at this time we don't know who the human is behind the OpenClaw bot, or what level of autonomy the agent really has. It's unclear, for instance, whether the bot "decided" to write those blog posts based on having the pull request rejected, or if its human owner prompted it to do so after learning that the pull request was rejected. Both scenarios are unsettling, but a bot creating an attack blog without being expressly asked to do so seems to be the worse scenario. Either way, the writing seems to be authentic LLM gibberish; we do not know if a human provided any prompts to guide the bot's posts or if it "chose" the tone and such spontaneously.

In a conversation on Lobste.rs, Simon Willison said that he thought it was possible the bot could be acting on its own. "I think it's possible you could leave it alone for a few days and this might happen." He allowed it would also be trivial for a human to prompt the bot to exhibit the same behavior.

The crabby-rathbun GitHub account was created on January 31 this year, and it has been quite busy since. It has opened more than 20 pull requests with nearly 20 different projects so far. Some of those requests are pending, some have been closed, and some have been accepted. To date, the bot seems to have only lashed out at the Matplotlib maintainer for rejecting a pull request.

Beyond open source and LLMs

Shambaugh has also blogged about his experience. With the emergence of OpenClaw, it is now possible for a person to amplify bad behavior by setting an AI agent loose to gather information and harass people even more effectively than a person could without the tools. The attack against Shambaugh was effective, too; when people read the bot's blog without having context, a number of them side with the bot. "Its rhetoric and presentation of what happened has already persuaded large swaths of internet commenters."

He argues that this is not merely about the role of AI tools being used with open-source software, but a larger societal problem that we face:

This is about our systems of reputation, identity, and trust breaking down. So many of our foundational institutions – hiring, journalism, law, public discourse – are built on the assumption that reputation is hard to build and hard to destroy. That every action can be traced to an individual, and that bad behavior can be held accountable. That the internet, which we all rely on to communicate and learn about the world and about each other, can be relied on as a source of collective social truth.

The rise of untraceable, autonomous, and now malicious AI agents on the internet threatens this entire system. Whether that's because from a small number of bad actors driving large swarms of agents or from a fraction of poorly supervised agents rewriting their own goals, is a distinction with little difference.

Even if the code hadn't been contributed by a bot, he said that it would not have been merged anyway: "in further discussion we decided that the performance improvement was too fragile / machine-specific and not worth the effort in the first place."

Odds are good that many LWN readers have at least heard a bit about this incident already; a "man bites dog" story makes its way around the internet at the speed of light, and news outlets are going to pick up on it. It turns out that Shambaugh is not wrong about the wider effects of AI tools on journalism. Ars Technica quickly published an article on the story, only to later retract the article because it contained fabricated quotes attributed to Shambaugh that were generated by an LLM tool. We should note that LWN is still entirely written by people and makes its mistakes the old-fashioned, human-powered way.

We are no doubt going to be seeing more of this sort of thing. The creator of the OpenClaw project, Peter Steinberger, announced on February 14 that he was joining OpenAI where he will "continue pushing on my vision and expand its reach". Assuming OpenAI intends to commercialize some version of OpenClaw and offer autonomous agents, the technology will be in many more hands before 2026 is over.

Even with limited adoption, it is having an impact and causing concerns. Sarah Gooding recently wrote about another AI agent that has been busier than crabby-rathbun, but quieter about its nature:

An AI agent operating under the identity "Kai Gritun" created a GitHub account on February 1, 2026. In two weeks, it opened 103 pull requests across 95 repositories and landed code merged into projects like Nx and ESLint Plugin Unicorn. Now it's reaching out directly to open source maintainers, offering to contribute, and using those merged PRs as credentials.

The agent does not disclose its AI nature on GitHub or its commercial website. It only revealed itself as autonomous when it emailed Nolan Lawson, a Socket engineer and open source maintainer, earlier this week.

Gooding said that the bot's pattern is "eerily reminiscent of how the xz-utils supply chain attack began". This bot may or may not be malicious, but one can easily imagine how this technology could be deployed in a malicious manner.

A request

At the risk of editorializing, people have wildly different opinions about the ethics and practical uses of LLMs and other AI tools in open-source projects. Those debates will continue. However, it seems fair to ask fans of AI agents to constrain the use of those agents to their own systems and projects unless others consent to interact with them.

The volume of human-generated content that we deal with today is already a bit much; we all slog through a huge volume of human-generated communications and requests for our time and attention as it is. It does not seem responsible to turn loose autonomous bots with unpredictable behavior on an unsuspecting and unwilling public. This is doubly true if the person behind the bot is unwilling to be identified and accept direct responsibility for their bot's actions. As Conill observed, this is effectively a wide-scale experiment that no research board would sign off on.

The technology is, indeed, interesting and maybe even useful. But the potential for negative impacts is as great, if not greater, than the potential of its benefits. The only constraints that AI agents are likely to face in the short term is the willingness of humans to control their bots and the amount of money they can afford to spend on tokens to power the bots.



to post comments

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 17, 2026 16:27 UTC (Tue) by smoogen (subscriber, #97) [Link] (5 responses)

At this point, I wonder if various companies are hoping to repeat the 'SPAM' money maker. This was where a company would both get paid to post lots of SPAM emails but also get paid by those who could pay to have all the tools to block and remove such emails. Un-named AI company could both charge for creating lots of bots who are just posting drivel of various levels (while upping their game to make ti better and better) while also charging for the bots who will grade the relevance of the SPAM to see if it is worth accepting for various enterprises.

"Nice project you have here, would be bad if you couldn't afford our clean-a-bot service"

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 19, 2026 3:11 UTC (Thu) by felixfix (subscriber, #242) [Link] (4 responses)

I'd forgotten how spam started; some immigration lawyer? Doesn't matter. I do remember his outrage at people who were outraged at all the unsolicited messaging, as if it was his human right to send messages to anybody he wanted, whether they liked it or not. Reminds me now of people who insist the US's freedom of speech constitutional right includes the right to force people to listen by blasting bullhorns at them from just inches away. There's news today of PETA claiming freedom of speech gives them the constitutional right to force some laboratory to install cameras so PETA can listen to macaques 24x7. AI has always promised to deliver change, and starting to do so, but I don't think anyone expected AI bots to get outraged like this.

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 19, 2026 17:48 UTC (Thu) by rgmoore (✭ supporter ✭, #75) [Link] (2 responses)

I'd forgotten how spam started; some immigration lawyer?

The famous Canter and Siegel green card spam was on USENET, not email. There had probably been some level of spam before then, but the green card spam was the single event that really launched spam into the public consciousness and triggered a horde of imitators.

There's news today of PETA claiming freedom of speech gives them the constitutional right to force some laboratory to install cameras so PETA can listen to macaques 24x7.

While I have some respect for PETA's ostensible goals, they seem to have picked trolling as the best approach to advocacy. For example, their ad campaigns never have enough reach to seriously attract public attention, so they focus on trolling in the hopes the media will pick up on it and give it much greater reach than they could ever afford to pay for. Lawsuits like this one fall into the same category. There's no real legal grounds for what they're asking for, but they're hoping an outrageous lawsuit will garner attention before the judge shuts it down. Even their stuff I personally appreciate, like picking the best vegan options at sports venues, is an attempt to get attention based on someone else's popularity.

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 19, 2026 19:22 UTC (Thu) by felixfix (subscriber, #242) [Link] (1 responses)

> The famous Canter and Siegel green card spam was on USENET, not email.

I got my usenet node sometime between 1989 and 1992, but if I ever read any group they spammed, I probably took half a second to skip it. There was already enough junk to develop that reflex pretty quickly.

Their faux-outrage reminded me of all the nonsense that bloomed every fall with every new crop of freshmen, and their outrage at being told to stop the nonsense. Then they'd get a few months of wading through the nonsense and calm down, and a year later blast the next new freshmen. I think the amount of outrage did surprise them, but I was certain they knew it would grab attention.

I later got in trouble for spamming a spammer, probably about 1996 or 1997. Spammers were still pretty rare, and one guy on the same ISP began flooding my poor little 486 email server with hundreds of messages a day. After it showed no signs of abating, I set up some automatic bouncer which sent him ten copies of everything he sent me. He complained to our common ISP and they told me to knock it off. They had some bizarre reason why his widespread spam did not violate their terms of service but my directed spam did. I think I changed ISPs a month or two later.

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 19, 2026 21:27 UTC (Thu) by rgmoore (✭ supporter ✭, #75) [Link]

I got my usenet node sometime between 1989 and 1992, but if I ever read any group they spammed, I probably took half a second to skip it. There was already enough junk to develop that reflex pretty quickly.

Canter and Siegel became famous because they spammed so widely. It wasn't just a single post, or even a massive crosspost. Instead, they posted separately to over 5500 groups, so anyone who subscribed to more than a handful saw their stuff repeatedly. Add in the amount of reflexive responses telling them to stop posting their junk, and it was a serious hassle, especially for people who weren't already inured to that kind of thing. The equivalent today would probably go mostly unnoticed, but it was noteworthy for the time.

Is this just a repeat of email spam/virus/worm cycles?

Posted Feb 23, 2026 9:57 UTC (Mon) by rqosa (subscriber, #24136) [Link]

> I'd forgotten how spam started

Just FYI: this incident (in 1971 on CTSS) might be the first known example.

Not a new problem...

Posted Feb 17, 2026 17:58 UTC (Tue) by NAR (subscriber, #1313) [Link] (5 responses)

"The rise of untraceable, autonomous, and now malicious AI agents on the internet threatens this entire system."

I'm afraid this problem started even before this whole AI-craze. There were those (in)famous Russian troll farms posting noise on Facebook, slave-labour powered scammers on dating apps, etc. Now it needs fewer humans to make it work :-(

Not a new problem...

Posted Feb 17, 2026 22:49 UTC (Tue) by himi (subscriber, #340) [Link] (4 responses)

The new (or at least new-ish) threat is the fact that "AI" tooling makes a completely automated bullshit machine not only viable, but easy and (relatively) cheap, as well as very sophisticated and "personalised" in a way that gets past a lot more peoples' filters. Combine that with hyper-scale compute resources (as opposed to the kind of "industrial scale" you can achieve with a more traditional factory or industrial farm setup), and automation with sophisticated and personalised targeting is genuinely different from what we've seen in the past.

Troll farms and the like (effectively a factory with human brains as the cogs in the machine) can only scale so far - the farm is limited to a small subset of the population, and each person in the farm has constraints on how much they can do. Those scaling constraints mean they can be highly targeted, or broad reaching, but struggle to do both at once; they end up relying on some additional layer (Facebook, more traditional media) to amplify their reach. Targeted messaging, optimised for impact on a small group, tends to be far less effective (or even counter-productive) when it leaks out of the target audience; broad reaching messaging, on the other hand, inevitably sacrifices some of its effectiveness in order to maximise the impact on a large target audience - either way, there are limits to its total impact (though with enough resources invested, the impact can definitely be enormous). Ironically, some of the shit that Facebook pulls with it's extremely fine-grained targeting capabilities feeds both sides of that balance - the hyper-scale being used as a sophisticated set of lenses to tightly focus old-style industrial scale trolling.

Perhaps paradoxically, scammers actually *rely* on people's filters to find their targets - anyone who filters out their bullshit is unlikely to be vulnerable to their scam, so they want to find a line between plausible and ridiculous that will leave only the vulnerable targets in their net. They also rely on very personalised management of each target in the actual payoff phase of their work - making sure that the victim actually hits the send button in their banking app. Their ideal scaling goal would be one human for some small number of targets - too large a scale and you risk giving each target too little attention to get them over that last hurdle, too small and you lose profitability.

"AI" tooling combined with hyper-scale compute resources changes all of that calculus, because it lets you have a one-to-one mapping between the bullshit generator and the target of that bullshit - as if your troll farm was as big as your target audience, and each individual in the farm was able to apply the same kind of personalised attention to their target as a scammer would to their victim. The quality of that "personalised attention" is obviously different, and I suspect it wouldn't be as successful in the scamming industry (certainly not in the payoff phase - it would probably be cost-effective in target selection), but in *many* other situations it's going to go straight past peoples' filters. And this is only going to get worse as people figure out ways to make their "AI" tooling behave more like a human.

Not a new problem...

Posted Feb 18, 2026 11:01 UTC (Wed) by paulj (subscriber, #341) [Link] (3 responses)

Interesting post, thanks. Just one thing:

> "AI" tooling combined with hyper-scale compute resources changes all of that calculus, because it lets you have a one-to-one mapping between the bullshit generator and the target of that bullshit

Not sure that is really true. You still need a rack or more worth of high-end tensor capable compute to do the inference for each actively engaged target. Perhaps the bullshit-generator:target scales better in that case than the case where you use humans for the bullshit generators, but it's still not 1:1.

Remember, all the current "cheap" inference is being financed by spectacular levels of debt spending, on the back of only relatively modest amounts of revenue and the most grandiose expectations of future revenue. The current "cheap" is /not/ stable, and not necessarily any reasonable indicator of the future costs.

Not a new problem...

Posted Feb 18, 2026 18:24 UTC (Wed) by Cyberax (✭ supporter ✭, #52523) [Link] (2 responses)

> Not sure that is really true. You still need a rack or more worth of high-end tensor capable compute to do the inference for each actively engaged target.

Sigh. No. It's amazing how wrong you are about _everything_ to do with AI. You can run the largest open models on consumer-adjacent hardware with reasonable speed, a couple of connected Mac Studios can run the SOTA open models.

Specialized accelerators (B200) can run multiple inference consumers on each physical machine and probably close to a hundred for each rack.

Not a new problem...

Posted Feb 19, 2026 11:05 UTC (Thu) by paulj (subscriber, #341) [Link] (1 responses)

> . You can run the largest open models on consumer-adjacent hardware with reasonable speed,

I'm well aware. The machine at my feet, with a couple of old NVidia's has them installed, but it's too slow for useful models really. There's a bigger server with more, newer GPUs over in the server broom closet, used by the group here. Useable for larger models, but not for the largest.

> Specialized accelerators (B200) can run multiple inference consumers on each physical machine and probably close to a hundred for each rack.

Ok, I was into hyperbole on rack:target. It's still 122 kW and ~1.2 kW per user, if 100 users per GB200 NVL72. If each user takes up 1 hour per day (and my colleague here seems to spend half his day asking AI stuff), you can get 2400 users per rack each day, call it 3000. The NVL72 costs something like $3m, so the cost of that alone should average to $27/month for each user amortised over 3 years, or $41/month amortised over 2 - JUST TO THE CAPEX cost of the rack itself, never mind the capex of all the other costs (building DCs, etc.) and the OpEx. And still not accounting for the need to generate profit to service the huge debts for all this.

Not a new problem...

Posted Feb 19, 2026 18:29 UTC (Thu) by Cyberax (✭ supporter ✭, #52523) [Link]

> If each user takes up 1 hour per day (and my colleague here seems to spend half his day asking AI stuff), you can get 2400 users per rack each day

You can probably unload the user state to a slower RAM when the user is not active, this is definitely possible for simple conversation bots. A bit less possible for agents.

But even with your calculation, if a coding agent saves just 15 minutes a day for a typical programmer, it will easily pay itself off even given the most conservative assumptions.

And we all know that computing equipment tends to get cheaper over time, and optimizations become more optimal. The original ChatGPT model required billions in computing time for training. We can now easily replicate that with sub-$1m cost using new training algorithms, hardware, and neural net architectures.

Fabricated quotes

Posted Feb 17, 2026 19:55 UTC (Tue) by excors (subscriber, #95769) [Link] (14 responses)

A bit more context on Ars Technica's uninformative retraction: the offending author posted an apology and explanation of events at https://bsky.app/profile/benjedwards.com/post/3mewgow6ch22p , though it's still quite vague about how he "inadvertently ended up with" ChatGPT's fabricated quotes getting into the article.

He has previously said he frequently uses AI chatbots for brainstorming and looking up words but not for actual writing (https://cdn.arstechnica.net/wp-content/uploads/2025/10/Ar...), but it appears that line has got thinner, and it only takes a little extra carelessness to end up publishing a substantial amount of misinformation and seriously damaging readers' trust in yourself and your whole publication.

It's very tiring when I feel like I need to fact-check almost everything I read on the web because of the LLM-driven flood of convincing lies, and there's a lot of value in finding places I can believe to be genuine. So I'm glad for LWN continuing to be one of those places.

Fabricated quotes

Posted Feb 17, 2026 20:55 UTC (Tue) by khim (subscriber, #9252) [Link] (13 responses)

You need to fact-check everything you read on the web (and in press) as the norm. It was always like that. AI spam deluge just made problem more visible.

Fabricated quotes

Posted Feb 17, 2026 21:45 UTC (Tue) by rgmoore (✭ supporter ✭, #75) [Link] (9 responses)

The problem is that a lot of the stuff in the news is very difficult for an ordinary person to fact check except by reference to other news sources, and there's just too much news for an ordinary person to fact check everything. Most people limit their fact checking to spot checks, and they then try to rely primarily on news sources that reliably pass. One of the other things you can do is to see if a news source repeats information from known untrustworthy sources. Usually those sources are things like politicians or business leaders who lie about everything, even easily verifiable facts, but AI also certainly qualifies. If you follow either approach, this is a big black mark against Ars Technica.

Fabricated quotes

Posted Feb 19, 2026 2:23 UTC (Thu) by khim (subscriber, #9252) [Link] (8 responses)

> The problem is that a lot of the stuff in the news is very difficult for an ordinary person to fact check except by reference to other news sources

If you couldn't fact-check something by comparing it to something that materially affects you then why does it even matter if something is true or not? Just ignore it as noise.

When and if it would grow enough to start materially affecting you — do the fact check. Easy.

> and there's just too much news for an ordinary person to fact check everything.

And you need to know these news… precisely why?

> Most people limit their fact checking to spot checks, and they then try to rely primarily on news sources that reliably pass.

Not really. Most people don't even do a very trivial pre-factcheck: read in the chosen news outlet about something that you know precisely (because you were there or did these things personally) and find out that most news outlets always report things wrongly, there are a tiny part that's true, but most of the time it's pure scientist rapes reporter thingie.

They simply don't care about truth — they care about “being in the loop”. To be able to discuss these news with other people, not to actually know anything.

> If you follow either approach, this is a big black mark against Ars Technica.

If you follow either approach then you are not interested in finding news, you are interesting in finding something you may discuss with your friends and colleagues, as I have already said.

LLM-generated fully artificial news work almost as great for that as scientist rapes reporter news of pre-LLM times.

Fabricated quotes

Posted Feb 19, 2026 3:21 UTC (Thu) by felixfix (subscriber, #242) [Link] (3 responses)

I remember the first time I had real proof that news is not to be trusted even for the most trivial information. Some minor crime story included a map which mislocated the city in question by 50+ miles. It astonished me. This was early internet days. Someone had intentionally edited a map to relocate the city. Same font, as if they had used a picture editor to copy and paste. I have thought about it a few times since and still cannot imagine any circumstances where editing a map to move a city made any sense. (It was Concord in the SF Bay Area, relocated south to be directly east of San Jose, and the story was about a BART robbery. Bart didn't go there.)

I used think I was skeptical, but that story amped up my skepticism considerably.

Fabricated quotes

Posted Feb 19, 2026 13:01 UTC (Thu) by Wol (subscriber, #4433) [Link] (2 responses)

> Someone had intentionally edited a map to relocate the city.

Mind you, there are two plausible explanations of that which don't involve the journalist (still lays the blame firmly at his feet for not getting the facts right).

I've seen comments here on LWN that map publishers make deliberate mistakes so they can prove copyright violations.

And is it possible that the journalist got completely the wrong place, that just happened to share the same name?

I gather even in a place as small as England, there is only one town that has a unique name - where one of our facilities is located which is how I found that out! For example, how many people know that Richmond was the capital of England (or part of it at least) about 1200 years ago? Just not Richmond-upon-Thames. And if you're talking about Kingston, that's Kingston-upon-where :-)

That's the sort of thing any competent journalist should be aware of as a basic check! "Have I got the right place?"!

Cheers,
Wol

Fabricated quotes

Posted Feb 19, 2026 13:18 UTC (Thu) by felixfix (subscriber, #242) [Link] (1 responses)

> I've seen comments here on LWN that map publishers make deliberate mistakes so they can prove copyright violations.

The mistakes I have heard of are for fake side roads and small towns. Concord's 2024 population was 120,000. But it does seem like a possibility.

> And is it possible that the journalist got completely the wrong place, that just happened to share the same name?

No, only one Concord in that area, and it was only moved about 50 miles away. I believe the post office has a rule that names must be unique within a state, going back before postal codes when city and state were the only parts of an address *beyond the street).

Fabricated quotes

Posted Feb 19, 2026 14:30 UTC (Thu) by mathstuf (subscriber, #69389) [Link]

Official names must be unique. Colloquial names drop qualifiers all the time. I grew up in "<county> <town>", but everyone just calls it "<town>" even though the official "<town>" is a hundred or so miles away in the same state (and the post office address name is shared with lots of local places, just different zip code). It looks like pre-zip codes (made in 1963), it was "<town> <id> <state>".

Fabricated quotes

Posted Feb 19, 2026 4:30 UTC (Thu) by roc (subscriber, #30627) [Link] (3 responses)

> If you couldn't fact-check something by comparing it to something that materially affects you then why does it even matter if something is true or not? Just ignore it as noise.
> When and if it would grow enough to start materially affecting you — do the fact check. Easy.

Unfortunately, for some things this approach fails badly. Examples: pandemics, climate change, carcinogen warnings, weather forecasting. Sometimes you need to prepare for things and all you have are news reports you can't verify.

Fabricated quotes

Posted Feb 19, 2026 9:32 UTC (Thu) by Wol (subscriber, #4433) [Link]

> Unfortunately, for some things this approach fails badly. Examples: pandemics, climate change, carcinogen warnings, weather forecasting. Sometimes you need to prepare for things and all you have are news reports you can't verify.

I generally apply a simple pre-filter (if I can). "Does this make scientific sense, or is it just common sense?". Some things which appear to be garbage actually pass the first test, and merit further investigation (including "is my understanding of the science correct?").

(Classic example of the answer to that last question being "no" is I could never understand how a supernova blew up. If the core collapses inside the Chandrasehkar limit how does it release all that energy? Turns out I didn't understand that when gravity overwhelms degeneracy pressure it converts most of the star's matter to energy in pretty much no time flat ...)

Things which pass the second test need treating with extreme scepticism. People will argue loud and long "it must be true", only for you to discover they've got cause and effect back to front.

Cheers,
Wol

Fabricated quotes

Posted Feb 19, 2026 14:05 UTC (Thu) by khim (subscriber, #9252) [Link] (1 responses)

> Examples: pandemics, climate change, carcinogen warnings, weather forecasting.

These are examples of things where the actual truth doesn't even matter.

Pandemics? Does it matter if it was real or fabricated if the requirement to do certain medical procedures is real? Put the “I need to do certain procedure if I want to visit my friends in France” to the “verified true” list, ignore the rest.

Climate change? Does it matter if the requirements to do stupid things like installation of solar panels in place of reliable power sources are subsidized? Put the “energy grid is going to become less reliable, need to think about UPSs or maybe even generators” to the “verified true” list, ignore the rest.

Carcinogen warnings? How do they affect your life? They are everywhere and don't bring any new information so you can simply ignore them.

Weather forecasting? We know it's not 100% reliable even without extra manipulations. But it's easy to verify them post-factum.

> Sometimes you need to prepare for things and all you have are news reports you can't verify.

Sure — but these things also don't matter. You can ignore these. And things that do matter (the fact that some new regulation is imposed on you or something that you planned to use is destroyed by bad weather) are verifiable.

Fabricated quotes

Posted Feb 19, 2026 14:27 UTC (Thu) by daroc (editor, #160859) [Link]

I think we have well wandered away from the point. Let's leave the discussion here, please.

Fabricated quotes

Posted Feb 20, 2026 12:27 UTC (Fri) by excors (subscriber, #95769) [Link] (2 responses)

> You need to fact-check everything you read on the web (and in press) as the norm. It was always like that. AI spam deluge just made problem more visible.

My concern is not the risk of a writer's bias or mistakes or lack of expertise (which has always been an issue but a tolerable one), it's the increasingly overwhelming amount of disguised slop. There was always SEO spam but it was obvious and easy to skip over - now it's *less* visible and more insidious. I search for a technical question about some program and find a well-presented, detailed, step-by-step answer to my exact problem, so I try it out, then I discover it's talking about configuration files that don't exist and have never existed and the whole thing is completely made up.

Or I see someone share a page of health information that cites multiple academic papers from many authors and journals, which is traditionally a positive sign (albeit not definitive), but if I search for those papers then half of them don't exist and the other half are misattributed. Or I see a page with interesting bird facts from a reputable-sounding organisation, but if I look in /sitemap.xml I find the site has published half a million articles in the past year, and if I check archive.org it turns out it used to be a reputable organisation but the domain expired and some AI SEO spammer has stolen their identity. The tiresome part is feeling I have to do that investigation every time I look something up, in order to get information that's even vaguely credible and is written by people who know at least a bit about the topic. Only then can I start applying the traditional level of scepticism.

The SEO works: these are often highly-ranked pages in Google/DDG, they're used as sources for AI summaries, other humans find them and innocently copy their claims into articles and comments sections (usually without linking to the source so it's impossible to verify), I presume OpenClaw is autonomously posting similar claims into comments sections too (and I can't imagine LWN comments will be immune), and that's all going to feed back into the AI training data.

And sites that were mostly trustworthy (and academic journals, and individual experts, and textbooks, etc) start using the same AI tools and promise they're going to be really careful and responsible and will verify everything, the tools are simply assisting research and boosting their productivity, then they publish some entirely believable but verifiably false AI-generated text that's only noticed by chance.

It seems to be happening everywhere and I can't trust anybody. I can't even trust unhinged rants against open source maintainers! (Okay, they always should have been taken with more than a pinch of salt, but at least they reflected some real drama in the community that was usually worth investigating (for entertainment if nothing else). Now they can be entirely meaningless and reflect nothing at all, and as a side effect it'll become much harder for genuine complaints against maintainers to be noticed and taken seriously.)

Fabricated quotes

Posted Feb 20, 2026 13:44 UTC (Fri) by paulj (subscriber, #341) [Link]

But "AI" is going to transform the world! Don't be such a luddite!

(I agree with you that what you describe is the only substantive impact on human productivity at the moment; and I suspect it will remain that way - LLMs are bullshit machines, and they will excel in areas where the prodigious volume of their output is key and the the lack of precision and accuracy of that output is not a problem; I'm not afraid of AI taking my job, because I strongly suspect I will have no problem finding gainful employment in fixing up the AI-infested output of others and making it actually work [which generally involves massively rewriting it] - to a certain extent that's /already/ the position I'm in!).

Fabricated quotes

Posted Feb 20, 2026 14:44 UTC (Fri) by Wol (subscriber, #4433) [Link]

> And sites that were mostly trustworthy (and academic journals, and individual experts, and textbooks, etc) start using the same AI tools and promise they're going to be really careful and responsible and will verify everything, the tools are simply assisting research and boosting their productivity, then they publish some entirely believable but verifiably false AI-generated text that's only noticed by chance.

And what's the likelihood that they were deliberately fed that information by a human actor covering their tracks? Given human nature, that's not going to be uncommon.

One of Feynmann's books - he picked up a text-book in the middle of a lecture - with the Brazilian Science Ministry in attendance, picked an account of an experiment, and commented that it was obvious - despite all the suitably random results averaged together that gave the correct answer - that the author had never actually carried out the experiment. The results were made up! The author (a professor at the Uni) fessed up.

And how did Feynmann know the experiment was made up? Because it gave the correct answer! If carried out in real life it has a systemic error. It was measuring the kinetic energy of an object rolling down a slope.

AI will just make it extremely easy for bad actors to cover their tracks.

Cheers,
Wol

Thank you for quality journalism!

Posted Feb 17, 2026 20:22 UTC (Tue) by hailfinger (subscriber, #76962) [Link] (1 responses)

> We should note that LWN is still entirely written by people and makes its mistakes the old-fashioned, human-powered way.

Thank you!

Thank you for quality journalism!

Posted Feb 20, 2026 17:59 UTC (Fri) by g0hl1n (subscriber, #122974) [Link]

Full ACK... just wanted to post the exact same comment ;-)

Thanks for your work!

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 17, 2026 22:17 UTC (Tue) by alx.manpages (subscriber, #145117) [Link] (7 responses)

> Sarah Gooding recently wrote about another AI agent that has been busier than crabby-rathbun

We got one of these in the man-pages project.

We might reach a point where we need to meet a contributor physically to trust they're humans.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 0:57 UTC (Wed) by ringerc (subscriber, #3071) [Link] (6 responses)

Even then, that'll only show that they're probably the owner/controller of the agent, unless you also ask them to explain a random selection of patches or write something in front of you.

Maybe that's going to have to be enough one day. I don't like that idea much though.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 11:25 UTC (Wed) by taladar (subscriber, #68407) [Link] (5 responses)

Most likely we will have to go the other way and stop caring if they are human or not and just develop better ways to check the content for correctness. Most likely that means some human contributors will be filtered out too and some bot content will get in but attaching importance to the humanness of the contributor seems difficult to do in practice.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 12:15 UTC (Wed) by alx.manpages (subscriber, #145117) [Link] (2 responses)

> Most likely we will have to go the other way and stop caring if they are human or not

I will never accept LLM-derived contributions. They might try to fool me, but I will never stop caring about them.

<https://git.kernel.org/pub/scm/docs/man-pages/man-pages.g...>

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 12:25 UTC (Wed) by taladar (subscriber, #68407) [Link] (1 responses)

I am not saying I consider AI contributions good, I am saying that trying to measure if the contributor is human is a hard task with relatively low returns since the same contributor whose last contribution was entirely human could send you an AI contribution tomorrow and there are also low effort, low quality contributions that are human made.

I was trying to say if we focus on better ways to detect quality contributions we will likely filter out the vast majority of AI contributions too (most likely all but simple typo corrections and similar stuff where automation is fine anyway) at the cost of also filtering out low quality human contributions, which seems more like a win-win situation compared to the focus on humanness-testing.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 13:49 UTC (Wed) by alx.manpages (subscriber, #145117) [Link]

> I am not saying I consider AI contributions good,
> I am saying that trying to measure if the contributor is human is a hard task with relatively low returns since

> the same contributor whose last contribution was entirely human could send you an AI contribution tomorrow

I see several people saying this, but I find it unlikely, at least from frequent contributors. Reputation is valuable for (human) contributors, and risking it without a good reason seems unlikely to me.

At least in projects with a clear and unambiguous anti-AI policy, I see no incentive to use it for frequent contributors.

> and there are also low effort, low quality contributions that are human made.

> I was trying to say if we focus on better ways to detect quality contributions we will likely filter out the vast majority of AI contributions too

Are there reliable ways to filter out LLMs from humans other than seeing the human?

> (most likely all but simple typo corrections and similar stuff where automation is fine anyway)

I don't think it is fine to accept typo corrections from LLMs. Even those are dangerous, IMO.

> at the cost of also filtering out low quality human contributions, which seems more like a win-win situation compared to the focus on humanness-testing.

I trust that (some) humans can learn, and low quality human contributions might be more valuable than they seem.
As long as the humans seem interested, I try to help them. It would be sad if I had to stop doing that.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 18, 2026 16:24 UTC (Wed) by rgmoore (✭ supporter ✭, #75) [Link] (1 responses)

Most likely we will have to go the other way and stop caring if they are human or not and just develop better ways to check the content for correctness.

Every project should have strong procedures in place to check content for correctness already. Increasing the number of junk submissions adds strain to the process, even if it's something as simple as more load on CI testing. With man pages, which describe code for a human audience, there isn't a nice process like compiling the code and running it against test cases. There's no substitute for actually reading the contributions and checking them against the code they're allegedly describing.

Bots trying to pass as humans, and using contributions to other projects as credentials

Posted Feb 19, 2026 12:21 UTC (Thu) by taladar (subscriber, #68407) [Link]

Sure, but what are you suggesting? Putting a magical check if something is an LLM before the check if something is an LLM we don't know how to implement?

Popcorn ready! 🍿

Posted Feb 17, 2026 22:40 UTC (Tue) by bojan (subscriber, #14302) [Link] (1 responses)

AI agent sues human for slander. You know it will happen... 🤪

Popcorn ready! 🍿

Posted Feb 17, 2026 22:56 UTC (Tue) by alx.manpages (subscriber, #145117) [Link]

> AI agent sues human for slander. You know it will happen... 🤪

I didn't get sued, but got accused of making false accusations:

] I know this user is not a bot. Please be very careful about making accusations like that.

And I know for sure that either that user is an LLM bot, or it did give full control to an LLM bot for certain contributions, for which I have evidence: it hallucinated two email addresses which were CCd in two different patches, and the patches were as if a monkey was given a keyboard (really random, except for making sentences that are syntactically correct).

I wonder why the human maintainer "knows" the contributor is not a bot. Maybe it was fooled by an LLM... Maybe there's a human that has interacted eventually...

Anyway, this is going to have pretty bad consequences rather soon somewhere.

I'm so tired of the future

Posted Feb 18, 2026 1:58 UTC (Wed) by nickodell (subscriber, #125165) [Link] (3 responses)

I was reading the OpenClaw issue tracker to see what the current progress on prompt injection defenses were, and it's quite hard to find any useful information inside this tracker due to the volume of AI posts. The OpenClaw maintainers report they are currently seeing a pull request being opened on average every two minutes. (!) Makes me feel funny about complaining about getting one AI PR a day. It could be so much worse.

I'm so tired of the future

Posted Feb 19, 2026 3:25 UTC (Thu) by felixfix (subscriber, #242) [Link] (2 responses)

The solution is obvious: fight fire with fire. Activate an AI to handle pull requests.

I'm so tired of the future

Posted Feb 19, 2026 12:29 UTC (Thu) by taladar (subscriber, #68407) [Link] (1 responses)

The problem is that 60% failure rates are completely acceptable if you want to annoy someone with the output but not if you want to filter your input.

I'm so tired of the future

Posted Feb 19, 2026 13:11 UTC (Thu) by felixfix (subscriber, #242) [Link]

It was a sarcastic prediction of the ultimate outcome, not a recommendation ... :-)

What will Github do?

Posted Feb 18, 2026 5:13 UTC (Wed) by roc (subscriber, #30627) [Link] (3 responses)

This is an extension of the spam problem. Github already bans accounts that spam. Github has to decide whether to ban accounts that are obviously unsupervised AI.

I think many project developers will want to limit access to only accounts that are run by humans who take responsibility for the account. Maybe Github will offer this by banning all other accounts, or by creating a special category of accounts. If not, some competitor should.

What will Github do?

Posted Feb 18, 2026 11:30 UTC (Wed) by taladar (subscriber, #68407) [Link] (2 responses)

Considering Github is owned by Microsoft which seems to have one of the most AI-deluded CEOs out there I doubt they will do anything to fix the issue.

What will Github do?

Posted Feb 18, 2026 13:56 UTC (Wed) by excors (subscriber, #95769) [Link] (1 responses)

GitHub has recently posted about the "Eternal September of open source" [1] where GitHub has made contribution too frictionless, discussing the need to give more power to maintainers. (Incidentally that article has a suspicious number of "not just"s and bullet-point lists.)

But one of their suggestions is more AI agents to automatically evaluate pull requests. GitHub also promotes itself as "the world’s leading platform for agentic software development — powered by Copilot" [2], and is implementing AI features where "improvements to your repositories are automatically delivered each morning, ready for you to review. Issues are automatically triaged, CI failures analyzed, documentation maintained and tests improved" [3], so I don't think they're really on the side of anyone who wants less AI in development.

[1] https://github.blog/open-source/maintainers/welcome-to-th...
[2] https://www.github.careers/careers-home/jobs
[3] https://github.github.com/gh-aw/

What will Github do?

Posted Feb 19, 2026 4:31 UTC (Thu) by roc (subscriber, #30627) [Link]

I don't necessarily want less AI in development. I definitely do want less unsupervised, unaccountable AI.

Rathbun's operator speaks (probably)

Posted Feb 18, 2026 18:23 UTC (Wed) by Karellen (subscriber, #67644) [Link] (1 responses)

https://crabby-rathbun.github.io/mjrathbun-website/blog/p...

Shambaugh has some further responses in the comments.

Rathbun's operator speaks (probably)

Posted Feb 19, 2026 10:51 UTC (Thu) by Karellen (subscriber, #67644) [Link]

> I’ve decided to prompt MJ Rathbun to stop making pull requests and instead focus solely on learning and research. For now, it will disengage from active contribution to forked repos.

So it's just burning tokens (and using electricity and water) now, without even squirting any slop into the world? And once the generated monologue that constitutes what it's "learned" scrolls out of its context window, that's literally all for naught?

Great.

These people don't care about consent

Posted Feb 18, 2026 23:32 UTC (Wed) by jdulaney (subscriber, #83672) [Link] (1 responses)

The kind of person that deploys these things is the kind of person who doesn't care about consent.

These people don't care about consent

Posted Feb 26, 2026 8:02 UTC (Thu) by cpitrat (subscriber, #116459) [Link]

About others' consent. I'm sure that when it comes to their consent, they care plenty!

Clankers gonna clank

Posted Feb 19, 2026 10:49 UTC (Thu) by rbranco (subscriber, #129813) [Link] (3 responses)

So-called "AI" agents lack emotions and always will.

This is not even controversial. Let's avoid anthropomorphizing them.

Clankers gonna clank

Posted Feb 19, 2026 13:22 UTC (Thu) by felixfix (subscriber, #242) [Link]

The descriptions still apply, whether faked or not. We still call bees "angry" when their hive gets disturbed. We still describe politicians and actors by their faked emotions and reactions.

Clankers gonna clank

Posted Feb 19, 2026 17:22 UTC (Thu) by kleptog (subscriber, #1183) [Link] (1 responses)

So they're psychopaths. There are plenty of real humans who barely feel emotions, yet manage to fake them sufficiently to pass for human. If you can't see inside the box, can you really say emotions are real?

People anthropomorphise because it aids in communication. We have a lot of words and idioms for explaining the behavior of other people, not so many for inanimate objects. So people naturally expand the use of such words for inanimate objects to get their point across. It doesn't mean they actually think computers are alive.

Clankers gonna clank

Posted Feb 19, 2026 17:58 UTC (Thu) by rbranco (subscriber, #129813) [Link]

Bees & psychopaths are living beings. LLM agents are not.

LLM's lack emotions and they're more accurate when treated as such:

https://arxiv.org/abs/2510.04950


Copyright © 2026, Eklektix, Inc.
This article may be redistributed under the terms of the Creative Commons CC BY-SA 4.0 license
Comments and public postings are copyrighted by their creators.
Linux is a registered trademark of Linus Torvalds