LWN: Comments on "On the use of LLM assistants for kernel development" https://lwn.net/Articles/1032612/ This is a special feed containing comments posted to the individual LWN article titled "On the use of LLM assistants for kernel development". en-us Fri, 17 Oct 2025 07:34:35 +0000 Fri, 17 Oct 2025 07:34:35 +0000 https://www.rssboard.org/rss-specification lwn@lwn.net Practical use of LLMs https://lwn.net/Articles/1035731/ https://lwn.net/Articles/1035731/ Wol <div class="FormattedComment"> <span class="QuotedText">&gt; LLMs are particularly effective for language-related tasks - obviously.</span><br> <p> <span class="QuotedText">&gt; LLMs are not so strong for programming,</span><br> <p> Programming is very much a language-related task. So which is it, LLMs are particularly effective for programming, or LLMs are useless at language? You can't have it both ways!<br> <p> And has been pointed out, LLMs are very capable at chucking out text that is simultaneously extremely plausible, and complete bullshit. THAT is the problem.<br> <p> The problem we want to solve isn't language, it's communication. And with absolutely no concept of comprehension or truth, LLMs are a serious liability.<br> <p> That said, LLMs are good at cleaning text up round the edges - until this eager-beaverness of all the peddlers of this rubbish actually gets seriously in the way of actually doing what you want to! I'm sick to death of Acrobat's desperation to "Let me summarise this document for you", when I'm actually looking for the *detail* I need which a summary will pretty much inevitably remove. The same with Sheets and Gemini - if I need detail to solve a problem, the LAST thing I need is Artificial Idiocy trying to summarise what I'm looking at!<br> <p> Cheers,<br> Wol<br> </div> Fri, 29 Aug 2025 15:04:31 +0000 Practical use of LLMs https://lwn.net/Articles/1035728/ https://lwn.net/Articles/1035728/ nim-nim <div class="FormattedComment"> Grammar checkers rely on a grammar, a formal description of language that has been checked and rechecked and proofed (sometimes at the state level), and is not legally-encumbered.<br> <p> LLMs are over-engineered plagiarism automatons that have no opinion on the correctness of the stuff they are plagiarising, except it should trigger strong reactions (because the field relies on advertiser money). It’s GIGO on a massive scale, with some checks added post-facto to limit the amount of garbage that spills out. No one has checked that every bit of content that has been used to train an LLM is correct, right, proper, good, free of legal encumbrances, etc.<br> <p> That’s the core difference and why LLM output requires human review.<br> </div> Fri, 29 Aug 2025 14:30:36 +0000 Generative-AI guidance from the Linux Foundation https://lwn.net/Articles/1033739/ https://lwn.net/Articles/1033739/ Wol <div class="FormattedComment"> <span class="QuotedText">&gt; Only material created by people can currently be copyrighted (in the USA at least).</span><br> <p> But the output of an LLM is based on the (copyrighted) material fed in. If the material that went in is copyrighted, saying "only material created by people ..." does not mean that what comes out of an LLM is copyright-free. All it means is that the LLM cannot add its own copyright to the mix.<br> <p> This is very clear in the European legislation, which says it's perfectly okay for an LLM to hoover up copyrighted material to learn from (exactly the same as a human would!), but makes no statement whatsoever as to whether the output is copyrightable or a derivative work (just like a human!)<br> <p> So assuming your statement is correct, US legislation says nothing whatsoever about whether the output of an LLM is copyrighted or not. All it says is that any *original* work by an LLM cannot be copyright.<br> <p> Cheers,<br> Wol<br> </div> Thu, 14 Aug 2025 14:10:58 +0000 Generative-AI guidance from the Linux Foundation https://lwn.net/Articles/1033711/ https://lwn.net/Articles/1033711/ rds <div class="FormattedComment"> The GPL is based upon copyright. Only material created by people can currently be copyrighted (in the USA at least). Using LLM's runs the risk of destroying the license of any code its used on. <br> <p> Disney just decided to not use machine generated images of an actor (Dwayne Johnson) because of concerns over the copyright status of the film.<br> </div> Thu, 14 Aug 2025 13:00:48 +0000 This is bad. https://lwn.net/Articles/1033662/ https://lwn.net/Articles/1033662/ mirabilos <div class="FormattedComment"> Why is Levin still a kernel developer? Why were his questionable commits not reverted? Why do Linux, Debian and OpenBSD not have a hard stance against any uses of the eso-fascist worldburning plagiarism machine like Gentoo and NetBSD do? Why do at least some of them let some of the slop in? Why is rsyslog still in Debian main?<br> <p> It’s a slippery slope starting by supporting slop sliding into the source.<br> </div> Wed, 13 Aug 2025 19:04:29 +0000 Practical use of LLMs https://lwn.net/Articles/1033478/ https://lwn.net/Articles/1033478/ pizza <div class="FormattedComment"> <span class="QuotedText">&gt; and it's going to be used to destroy trillions of dollars of human capital in the meantime, at a time when we probably cannot afford it.</span><br> <p> To me it's not the "destruction" of so much [human] capital but the wasted/squandered opportunities.<br> <p> <p> <p> </div> Tue, 12 Aug 2025 17:21:59 +0000 Practical use of LLMs https://lwn.net/Articles/1033476/ https://lwn.net/Articles/1033476/ raven667 <div class="FormattedComment"> I was saying on Mastodon the other day that when the OpenAI board tried to oust Sam Altman after he offered ChatGPT to the public that was probably the right call, they already had tested and knew the weaknesses of LLM models as a path to "AGI", but once the hype-train had left the station it proved impossible to stop. It's unlikely that any of this "AI" fever will work out to our betterment in the long run, and it's going to be used to destroy trillions of dollars of human capital in the meantime, at a time when we probably cannot afford it.<br> </div> Tue, 12 Aug 2025 17:16:22 +0000 Practical use of LLMs https://lwn.net/Articles/1033410/ https://lwn.net/Articles/1033410/ wtarreau <div class="FormattedComment"> Absolutely!<br> <p> But in addition starting to be careful about LLMs also teaches people to be careful of other people looking too smart. There is a huge confusion between knowledge and intelligence in general. Lots of people use the term "smart" or "intelligent" to describe a very knowledgeable person, and consider that someone lacking culture is "dumb". But I've seen people who, once explained the details of a problem, would suggest excellent ideas on how to solve them. *This* is intelligence. Those who only know everything and cannot use it except to look smart in conversations are just parrots. Of course it's way better when you have the two at once in the same person, and often smart people like to learn a lot of new stuff. But each profile has its uses. Right now LLMs solve only one part of the deduction needed for intelligence, and know a little bit of everything but nothing deeply enough to express a valid opinion or advice. Yes most people (as you say, 99.999% to stick with this thread) tend to ask them advices and opinions on stuff they are expected to know well since coming from the internet, but that they only superficially know.<br> </div> Tue, 12 Aug 2025 11:48:46 +0000 Generative-AI guidance from the Linux Foundation https://lwn.net/Articles/1033239/ https://lwn.net/Articles/1033239/ cesarb <div class="FormattedComment"> <span class="QuotedText">&gt; I though pretty much all the LLMs these days place additional restriction(s?), in particular, that you cannot use the output to improve another LLM.</span><br> <p> Not all of them. For instance, Qwen3 (<a href="https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507">https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507</a> and others) uses the Apache 2.0 license, and DeepSeek-R1 (<a href="https://huggingface.co/deepseek-ai/DeepSeek-R1">https://huggingface.co/deepseek-ai/DeepSeek-R1</a>) uses the MIT license (though see the note in that page about its distilled variants, the license depends on the model used as the base).<br> </div> Mon, 11 Aug 2025 13:12:24 +0000 Practical use of LLMs https://lwn.net/Articles/1033229/ https://lwn.net/Articles/1033229/ paulj <div class="FormattedComment"> <span class="QuotedText">&gt; I'm well aware that you cannot expect from an LLM to be precise/exact, because it should not be seen as a sequential computer program that can be debugged and made reliable, </span><br> <p> The problem is (invent plausible stat and confidently handwave it about - highly appropriate in a thread on LLMs!) 99.999% of the populace doesn't know this, and lack the combination of technical background, curiosity and time to come to understand this. They think - because the hype machine (the effective combination of companies wanting to sell stuff and non-technical media jumping on pushing the latest buzz) has told them so - that this stuff is "intelligent" and will solve all problems.<br> </div> Mon, 11 Aug 2025 10:11:43 +0000 Practical use of LLMs https://lwn.net/Articles/1033228/ https://lwn.net/Articles/1033228/ paulj <div class="FormattedComment"> Asimov has (at least) 1 story around this. E.g. Bicentennial Man (later made into a film, with Robbin Williams as the robot Andrew).<br> </div> Mon, 11 Aug 2025 10:06:19 +0000 Don't Ask, Don't Tell https://lwn.net/Articles/1033213/ https://lwn.net/Articles/1033213/ kleptog <div class="FormattedComment"> <span class="QuotedText">&gt; Having seen some AI code (that I was given) I wasn't impressed. It did the job, but it wasn't what I would have expected from someone who knew our coding style.</span><br> <p> But then it's easy right? "Doesn't match our coding style" is a perfectly valid reason to reject a patch.<br> <p> I believe I got it from the PostgreSQL lists: after your patch the code should look like it's always been there.<br> <p> Arguably, if new code doesn't follow the coding style (which is much broader than just where to put whitespace) then the author has not yet understood the code we'll enough to be submitting. Which covers the LLM case perfectly.<br> </div> Mon, 11 Aug 2025 07:47:05 +0000 Don't Ask, Don't Tell https://lwn.net/Articles/1033163/ https://lwn.net/Articles/1033163/ Wol <div class="FormattedComment"> <span class="QuotedText">&gt; How they came up with the code, what tools they used, might be interesting, but not more than what school they went to or what other projects they are working on. It's tangential in the end.</span><br> <p> Not if it affects the TYPE of bug that is in the code! As I think someone else pointed out, AIs and humans make different sorts of bugs. And if you don't know whether it was an AI or a human, it either (a) makes review much harder, or (b) makes missing things much more likely.<br> <p> Having seen some AI code (that I was given) I wasn't impressed. It did the job, but it wasn't what I would have expected from someone who knew our coding style. <br> <p> At the end of the day, I'm all for "no surprises". Who cares if it's an AI or a person. What matters is that it's declared, so the next guy knows what he's getting.<br> <p> Cheers,<br> Wol<br> </div> Sun, 10 Aug 2025 14:45:39 +0000 Don't Ask, Don't Tell https://lwn.net/Articles/1033159/ https://lwn.net/Articles/1033159/ abelloni <div class="FormattedComment"> This is not true, there is a difference between a patch made to silence a static checker and one for an issue that was actually seen in the field.<br> </div> Sun, 10 Aug 2025 11:16:42 +0000 AI generated code is not useful https://lwn.net/Articles/1033158/ https://lwn.net/Articles/1033158/ alx.manpages <div class="FormattedComment"> "Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it."<br> <p> — Brian W. Kernighan and P. J. Plauger in The Elements of Programming Style.<br> <p> The people defending that LLMs might make it easy for new programmers (which are otherwise unable to contribute code) to contribute code, somehow expect those new programmers to be able to review the code produced by an LLM?<br> <p> And for people that are already good programmers, will this reduce the work? Or will it increase it? <br> <p> You've changed the task of authoring code --in which case you often self-restrict to a set of coding standards that significantly reduce the possibility of bugs--, to the task of reviewing code --which by nature is already twice as hard--, and fully unrestricted, because you can't trust an LLM to consistently self-restrict to some rules. The bugs will appear in the most unexpected corners.<br> <p> Even for reviewing my own code, I wouldn't use an LLM. Reason: it might let two bugs pass for each one it catches, and I might have a false feeling of safety. I prefer knowing the limits of my deterministic tools, and improve them. And finding quality reviewers. That's what it takes for having good code.<br> <p> Abandon all hope, ye who accept LLM code.<br> </div> Sun, 10 Aug 2025 11:05:50 +0000 Practical use of LLMs https://lwn.net/Articles/1033089/ https://lwn.net/Articles/1033089/ khim <font class="QuotedText">&gt; The brain has a lot of dedicated hardware, be it visual recognition, auditory recognition, whatever.</font> <p>Isn't that GPT-5 “tools” and voice recognition in <a href="https://gemini.google/overview/gemini-live/">Gemini Live</a> is for?</p> <font class="QuotedText">&gt; AI runs on pure general purpose hardware.</font> <p>Not really. It can be run, in theory, on general purpose hardware, but it's not clear if GPT-5 run on general purpose hardware would be at all practical.</p> <p>Even if you just think about <a href="https://en.wikipedia.org/wiki/Bfloat16_floating-point_format">BF16</a>… it's pretty specialized thingie.</p> <font class="QuotedText">&gt; And has been pointed out, a lot of the brain's special-purpose hardware is survival-ware - if the hardware gets it wrong, it's likely to end up as a lion's lunch, or whatever ...</font> <p>Sure, but do we actually use that hardware where we are writing code? Somehow I doubt it. It's like arguing that LLM couldn't write good code because it doesn't have liver… sure, liver is very important for human, but lack of liver is not what stops LLM from being good software designer.</p> Sat, 09 Aug 2025 10:40:24 +0000 Practical use of LLMs https://lwn.net/Articles/1033088/ https://lwn.net/Articles/1033088/ Wol <div class="FormattedComment"> <span class="QuotedText">&gt; That's not imitation of human brain, though. That's imitation of insect brain or, maybe, a chimps brain (although a chimps have world model even if they are less complicated than humans world model). It's pure reaction with nothing to control the “train of though” and to stop it from derailing.</span><br> <p> It's not an imitation of ANY brain. Think about it. The brain has a lot of dedicated hardware, be it visual recognition, auditory recognition, whatever. And a small veneer of general purpose hardware over the top. AI runs on pure general purpose hardware.<br> <p> And has been pointed out, a lot of the brain's special-purpose hardware is survival-ware - if the hardware gets it wrong, it's likely to end up as a lion's lunch, or whatever ...<br> <p> Cheers,<br> Wol<br> </div> Sat, 09 Aug 2025 10:32:04 +0000 Practical use of LLMs https://lwn.net/Articles/1033084/ https://lwn.net/Articles/1033084/ khim <font class="QuotedText">&gt; I.e. in order to think like us they have to be as unreliable.</font> <p>Nope. It order to think you need <b>reliable</b> world model somewhere under all these words. Half-century old <a href="https://en.wikipedia.org/wiki/SHRDLU">SHRDLU</a> may think while ChatGPT-5 couldn't.</p> <p>Sure, humans make mistakes (especially when they are distracted), but they may also notice them automatically and fix them. Thus doesn't work with LLMs, in fact if you try to push them they become <a href="https://economictimes.indiatimes.com/tech/artificial-intelligence/smarter-but-less-accurate-chatgpts-hallucination-conundrum/articleshow/120432220.cms">even less accurate, then when they are not “thinking”</a>.</p> <font class="QuotedText">&gt; as something trying to imitate our brain with many inter-connections based on what was learned, probabilities and noise</font> <p>That's not imitation of human brain, though. That's imitation of insect brain or, maybe, a chimps brain (although a chimps have world model even if they are less complicated than humans world model). It's pure reaction with nothing to control the “train of though” and to stop it from derailing.</p> <p>The best illustration to what is happening with “reasoning” LLMs is <a href="https://en.wikipedia.org/wiki/Taylor_series">picture from Wikipedia in the article Taylor series</a> where it shows “<code>sin x</code> and its Taylor approximations by polynomials of degree 1, 3, 5, 7, 9, 11, and 13 at x = 0”.</p> <p>It's very easy to see that as “as the degree of the Taylor polynomial rises, it approaches the correct function” – but if you actually look on picture you'll notice <b>how</b> it does that: it becomes <b>ever more precise</b> in the small, but growing area around zero, but, simultaneously, also become <b>ever more absurdly wrong</b> in the area around that central part.</p> <p>And that's what is happening with LLMs: they are becoming ever more impressive at “one-shotting” things, yet. simultaneously, ever more helpless with attempts to handle a long series of tasks.</p> <p>This is similar to how very small kids behave, but eventually human learns to double-check and self-verify things… LLMs couldn't learn that, they simply have no mechanisms suitable for that.</p> <p>The latest fad in AI is to attach “tools” to LLMs and hope that Python interpreter would be able to work a reliable replacement for a world model. It wouldn't work: this would slightly expand area where LLMs would be able to “one-shot” things, but wouldn't fix the fundamental flaw in their construction.</p> Sat, 09 Aug 2025 09:42:54 +0000 Generative-AI guidance from the Linux Foundation https://lwn.net/Articles/1033086/ https://lwn.net/Articles/1033086/ jepsis <div class="FormattedComment"> <span class="QuotedText">&gt; I basically see only two possibilities. Either the submitter just cannot license the code as GPL-2.0 due to the restriction above, or they are risking their subscription to the LLM with every patch submission.</span><br> <p> Patches to the Linux upstream are always derivative works of Linux and therefore fall under the GPLv2. In most cases, authors of patches or patch sets to Linux cannot claim separate copyright, and they typically do not meet the threshold of originality. Using any tools or AI does not change this.<br> <p> Of course, if someone submits an entirely new and unconventional file system like 'VibeFS', copyright issues might arise. However, it is still highly unlikely that such a contribution would be approved, regardless of the tools used.<br> </div> Sat, 09 Aug 2025 09:39:21 +0000 Practical use of LLMs https://lwn.net/Articles/1033081/ https://lwn.net/Articles/1033081/ excors <div class="FormattedComment"> <span class="QuotedText">&gt; Their value however is in having abilities to directly use computer-based tools without having to physically move fingers, so they can use calculators and web search faster than us.</span><br> <p> One example I've seen is giving ChatGPT 5 - which was announced as having "PhD-level intelligence" - the prompt "Solve: 5.9 = x + 5.11". When I repeated it myself, 50% of the time it said x=0.79, and 50% of the time it said x=-0.21.<br> <p> In both cases it gave a superficially reasonable step-by-step explanation, saying things like "5.90 - 5.11 = (5.90 - 5.11) = 0.79 but since 5.90 is less than 5.11 in the hundredths place, the result will be negative: x = -0.21". That's nonsense, but it's confidently-stated half-correct nonsense, which engenders undeserved levels of trust.<br> <p> In theory the system could make use of external calculators and powerful, reliable, human-designed algebraic tools. In practice it doesn't - it does probabilistic calculations on language tokens, resulting in something that sounds like a mathematical calculation but actually isn't, making it untrustworthy for even trivial tasks like this. (And somehow this is worth half a trillion dollars.)<br> </div> Sat, 09 Aug 2025 09:09:23 +0000 Generative-AI guidance from the Linux Foundation https://lwn.net/Articles/1033077/ https://lwn.net/Articles/1033077/ gray_-_wolf <div class="FormattedComment"> <span class="QuotedText">&gt; In short, this guidance suggest that developers should ensure that the tool itself does not place restrictions on the code it generates,</span><br> <p> This is interesting. I though pretty much all the LLMs these days place additional restriction(s?), in particular, that you cannot use the output to improve another LLM. Ignoring the hypocrisy of slurping all of GitHub and then putting this rule on their products, how does that work with submitting the code to the kernel?<br> <p> I basically see only two possibilities. Either the submitter just cannot license the code as GPL-2.0 due to the restriction above, or they are risking their subscription to the LLM with every patch submission.<br> <p> What am I missing here?<br> <p> <span class="QuotedText">&gt; and that said code does not incorporate any pre-existing, copyrighted material.</span><br> <p> And how exactly am I supposed to ensure this?<br> </div> Sat, 09 Aug 2025 07:45:51 +0000 Practical use of LLMs https://lwn.net/Articles/1033074/ https://lwn.net/Articles/1033074/ wtarreau <div class="FormattedComment"> I don't know how robots behave in movies (I don't have a TV) but I'm well aware that you cannot expect from an LLM to be precise/exact, because it should not be seen as a sequential computer program that can be debugged and made reliable, but as something trying to imitate our brain with many inter-connections based on what was learned, probabilities and noise. I.e. in order to think like us they have to be as unreliable. Their value however is in having abilities to directly use computer-based tools without having to physically move fingers, so they can use calculators and web search faster than us. But the risk of inaccurately recopying a result remains non-null... like with humans who can get distracted as well.<br> </div> Sat, 09 Aug 2025 05:57:20 +0000 Practical use of LLMs https://lwn.net/Articles/1033069/ https://lwn.net/Articles/1033069/ Wol <div class="FormattedComment"> <span class="QuotedText">&gt; what they lack are “common sense”</span><br> <p> Given that "common sense" isn't common, and rarely makes sense, maybe that's just as well!!!<br> <p> Cheers,<br> Wol<br> </div> Fri, 08 Aug 2025 22:41:40 +0000 Practical use of LLMs https://lwn.net/Articles/1033063/ https://lwn.net/Articles/1033063/ khim <font class="QuotedText">&gt; The main problem with "matching humans" is that they'll have to pass by empathy, empathy, self-conciousness and some may even develop their own religions etc.</font> <p>Surprisingly enough that's already covered. Existing chatbots don't have “empathy”, “empathy” or “self-conciousness”, but they imitate them well enough to <a href="https://www.reuters.com/sustainability/boards-policy-regulation/google-ai-firm-must-face-lawsuit-filed-by-mother-over-suicide-son-us-court-says-2025-05-21/">achieve pretty disturbing results</a>. And I'm pretty sure they would do a superb job working as missionaries for various religious sects. No problem there at all: nefarious uses of LLMs scale surprisingly well.</p> <p>LLMs fail utterly when long chains of logical reasoning is needed, though.</p> <font class="QuotedText">&gt; So in the end these assistants will end up being new workers with all the same limitations and problems as other ones in the real world and will not solve that many issues for enterprises, except that they'll eat more power ;-)</font> <p>Highly unlikely. In fact the biggest obstacle to the use of LLMs is the fact that people try to apply what they have learned from books and movies about how “sentient robots” would behave over last century or so. Which is understandable, but also incredibly wrong.</p> <p>In books and movies “sentient robots” are always logical, correct and precise and it's a big problem for them to express emotion or simulate empathy… in real world LLMs can do all these things that “sentient robots” from all these countless books and movies struggled pretty easily… what they <b>couldn't do</b> are things that people expect them to do: logical reasoning, precision, reproducibility…</p> <p>That's another thing that plagues the whole industry: what all the presentations and demos portray and “sell” and what CEO expect to buy… are these “sentient robots” from movies. What they get… is something entirely different, something totally unsuitable for the role where “sentient robots” would fit perfectly.</p> <p>That's why <a href="https://gizmodo.com/klarna-hiring-back-human-help-after-going-all-in-on-ai-2000600767">Klarna rehires people back</a>, and IBM hires people for <a href="https://www.wsj.com/articles/ibm-ceo-says-ai-has-replaced-hundreds-of-workers-but-created-new-programming-sales-jobs-54ea6b58">“critical thinking” focused domains</a> and Duolingo <a href="https://fortune.com/2025/05/24/duolingo-ai-first-employees-ceo-luis-von-ahn/">puts people back</a>… it's all because LLMs are the total opposite from “sentient robots” in movies.</p> <p>If you read a summary papers written by press-people then you would hear how people are rehired because robots “lack empathy” or “emotions”, but that's a big fat lie: robots have more than enough empathy and emotions, spammers just simply <b>love</b> that aspect of LLMs… what they lack are “common sense” and “logic”.</p> Fri, 08 Aug 2025 21:09:44 +0000 Practical use of LLMs https://lwn.net/Articles/1033062/ https://lwn.net/Articles/1033062/ wtarreau <div class="FormattedComment"> The main problem with "matching humans" is that they'll have to pass by empathy, emotions, self-conciousness and some may even develop their own religions etc. Then at this point we'll have laws explaining how it's inhumane to treat an AI assistant badly by making it work endlessly and ignoring its suffering. So in the end these assistants will end up being new workers with all the same limitations and problems as other ones in the real world and will not solve that many issues for enterprises, except that they'll eat more power ;-)<br> </div> Fri, 08 Aug 2025 20:45:26 +0000 Practical use of LLMs https://lwn.net/Articles/1033041/ https://lwn.net/Articles/1033041/ khim <font class="QuotedText">&gt; Not really. LLMs don't deal in facts, they deal in probabilities, as in “what word is most likely to complete this partial sentence/paragraph/text?”</font> <p>Yes, but that's entirely different kettle of fish: humans have the world models, in fact they start form in human brain before humans learn to speak, starting from <a href="https://en.wikipedia.org/wiki/Peekaboo">peekaboo</a> and <a href="https://en.wikipedia.org/wiki/Hide-and-seek">hide-and-seek</a> games.</p> <p>LLMs doesn't have anything remotely similar to that, that's why they couldn't say “I don't know how to do that”: humans say that when their world model shows the “hole” and LLMs couldn't do that since there are no world model, it's all probabilities all the way down.</p> <p>In rare cases where says “I don't know” or “this thing is probably doesn't exist” (it <a href="https://www.os2museum.com/wp/ai-responses-may-include-mistakes/">happens sometimes, if rarely</a>) they simply found it highly probably, based on their training set, that this response would be, probably, the most appropriate one.</p> <p>The only “memory” LLM have are related to probabilities… that doesn't mean that there are long-term memory, it just means that it's different from what humans have.</p> <font class="QuotedText">&gt; If you ask Sam Altman he will tell you that OpenAI is only a year or so (and a few tens of billions of dollars) away from “artificial general intelligence”, but he's been doing that for years now and it's very hard to see how that would work given what they've been doing so far.</font> <p>That's yet another kettle of fish. I think AGI would soon be relegated to the annals of history, anyway: it's obvious that pure scaling wouldn't give anything similar to “average human worker” any time soon – and AGI is something that have marketable appeal <b>only</b> in that “scale is all you need” world.</p> <p>If we would be forced to match human capabilities by slowly and painstakingly adding more and more specialized modules, then AGI loses it's appeal: is still achievable – but somewhere in XXII or XXIII century where last 0.01% of something that human was doing better than pre-AGI system is conquered.</p> <p>By that time our AI is so drastically superhuman <b>at everything else</b> that saying that we have reached AGI no longer makes sense. It's more of “ oh yeah, finally… at arrived… what else is new?” moment, rather that something to talk about.</p> Fri, 08 Aug 2025 16:08:41 +0000 Practical use of LLMs https://lwn.net/Articles/1033038/ https://lwn.net/Articles/1033038/ anselm <blockquote><em>when LLM model is trained it “remembers” facts its “long-term memory”.</em></blockquote> <p> Not really. LLMs don't deal in facts, they deal in probabilities, as in “what word is most likely to complete this partial sentence/paragraph/text?” These probabilities can be skewed in various ways through training and prompting, but it is important to keep in mind that to an LLM, the world is essentially alphabet soup – it has no underlying body of abstract factual knowledge from which it could draw logical conclusions like we humans do. </p> <p> LLMs can certainly produce results that seem impressive, but in the long run they're probably just a side branch on the path to actual AI. If you ask Sam Altman he will tell you that OpenAI is only a year or so (and a few tens of billions of dollars) away from “artificial general intelligence”, but he's been doing that for years now and it's very hard to see how that would work given what they've been doing so far. </p> Fri, 08 Aug 2025 15:52:59 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1033039/ https://lwn.net/Articles/1033039/ laurent.pinchart <div class="FormattedComment"> <span class="QuotedText">&gt; LLMs do precisely the opposite: they make first ever patch look better than your average patch, but they make it harder for a newcomer to “eventually graduate into "real" kernel development”.</span><br> <p> An interesting study on that topic: "Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task" (<a href="https://arxiv.org/abs/2506.08872">https://arxiv.org/abs/2506.08872</a>)<br> </div> Fri, 08 Aug 2025 15:46:39 +0000 Practical use of LLMs https://lwn.net/Articles/1033024/ https://lwn.net/Articles/1033024/ khim <font class="QuotedText">&gt; In addition I'm pretty sure we'll start to imitate the way we currently function with short-term and long-term memory with conversion phases that we call "sleep" in our cases.</font> <p>That's already the case: when LLM model is trained it “remembers” facts its “long-term memory”.</p> <p>The only problem: it's entirely cost-prohibitively to offer mode where one may train LLM and later its long-term memory.</p> <p>In fact it's even prohibitively expensive to run existing models, too, thus pretty soon we would see serious <b>regressions</b> in all these tools capabilities.</p> <p>An era of “AI assistants” that you describe would come 10-20 years down the road, maybe even later.</p> Fri, 08 Aug 2025 15:17:20 +0000 Practical use of LLMs https://lwn.net/Articles/1033016/ https://lwn.net/Articles/1033016/ wtarreau <div class="FormattedComment"> <span class="QuotedText">&gt; You can not teach LLM anything.</span><br> <p> Note that I purposely said "AI assistants" not "LLM". LLMs are dumb because they're only based on language and I anticipate that future generations will come leveraging more of the multi-modal approach and will be more flexible like humans, by involving multiple senses at once. In addition I'm pretty sure we'll start to imitate the way we currently function with short-term and long-term memory with conversion phases that we call "sleep" in our cases. LLMs can already do quite impressive things and that's unfortunately why people believe they're very smart. But they can be impressively limited and dumb as well sometimes. Note that to be honest we all know super dumb humans with which we gave up trying to explain certain things. I fear the day we'll have to run interviews to AI assistants to decide if they're smart enough to be hired...<br> </div> Fri, 08 Aug 2025 14:57:50 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032946/ https://lwn.net/Articles/1032946/ kleptog <div class="FormattedComment"> One of the most promising developments I see for code LLMs is the switch to a more diffusion style. So rather than generating a sequence of words, it's working on a grid of characters like your editor and iterating over that in the same way that stable-diffusion produces images. This means it has a chance of "seeing" the code as a whole and making changes globally. And so noticing mismatches between different parts and aligning them in the next iteration. So more like how people code.<br> <p> The step after that would be an LLM iterating over an AST so it doesn't have to worry about getting the syntax right, but I haven't read about that yet. It's not clear to me if that technology even exists yet.<br> </div> Fri, 08 Aug 2025 12:33:32 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032941/ https://lwn.net/Articles/1032941/ jepsis <i>That's not “ first-time user working with LLVM” (LLM, I assume?). That's “experienced kernel developer trying LLM”.</i><p> Sure. Good example. It would have been good to have that sentence checked by AI, as it would likely have corrected it. Fri, 08 Aug 2025 09:43:48 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032938/ https://lwn.net/Articles/1032938/ khim <p>That's not “ first-time user working with LLVM” (LLM, I assume?). That's “experienced kernel developer trying LLM”.</p> <p>First time user request would be more of “here's the spec for that hardware that I have, write driver to it”. And then the resulting mess is sent to maintainer, warts, bugs and all.</p> Fri, 08 Aug 2025 09:32:02 +0000 Don't Ask, Don't Tell https://lwn.net/Articles/1032937/ https://lwn.net/Articles/1032937/ rgb <div class="FormattedComment"> At this point in time I think the only reasonable stance is a don't ask, don't tell policy, which is also basically the status quo anyhow.<br> At the end of the day, a human is the author of the patch. He or she is responsible for the content and also the point of trust that can hold or break.<br> How they came up with the code, what tools they used, might be interesting, but not more than what school they went to or what other projects they are working on. It's tangential in the end.<br> </div> Fri, 08 Aug 2025 09:15:39 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032933/ https://lwn.net/Articles/1032933/ jepsis I don’t see any issue with a first-time user working with LLVM.<p> Here are some examples of useful prompts:<p> <i>Is the naming of functions and variables consistent in this subsystem?<p> Are the comments sufficient, or should they be added to or improved?<p> If I were to submit this upstream, what aspects might attract nitpicking?<p> Does the commit message accurately reflect the commit, or are there any gaps?</i> Fri, 08 Aug 2025 09:04:41 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032931/ https://lwn.net/Articles/1032931/ khim <font class="QuotedText">&gt; I'd argue: *If* this lowers the bar and makes more people eventually graduate into "real" kernel development, that's not purely negative.</font> <p>LLMs do precisely the opposite: they make first ever patch look better than your average patch, but they make it <b>harder</b> for a newcomer to “eventually graduate into "real" kernel development”.</p> <p>That's precisely the issue with current AI: degradation of output. LLMs don't have a world model and when you try to “teach” them they start performing worse and worse. To compensate their makers feed them terabytes, then petabytes of himan-produced data… but that well is almost exhausted, there are simply no data to feed into these. And this scaling only improves the initial output, it does nothing to the lack of the world model and ability to learn during the dialogue.</p> <p>Worse: as <a href="https://en.wikipedia.org/wiki/Winthrop_Kellogg#The_Ape_and_the_Child">we know</a> than when ape and human interact human turns into ape, not the other way around.</p> <p>The chances are high that story with LLMs would be the same: when complete novices would try to use LLMs to “become a kernel developers” they would become more and more accepting to LLM flaws instead of learning to fix them. This, too, would increase load placed on maintainers.</p> <font class="QuotedText">&gt; LLMs are still improving fast.</font> <p>Yes and no. They are feed more and more data, which improves the initial response, but does nothing to gradual degradation of output when you try to improve it.</p> <p>Sooner or later you hit the “model collapse” threshold and then you have to start from scratch.</p> <font class="QuotedText">&gt; but also pretty good at knowing what they are not good at.</font> <p>So far that haven't worked at all. LLMs are all too happy to generate nonsense output instead of admitting that they don't know how to do something.</p> <font class="QuotedText">&gt; maybe make an LLM review patches where LLM contributed</font> <p>Given the fact that LLMs tend to collapse when feed their own input (that's why even most expensive plans don't give you the ability to generate long outputs, instead they give you the ability to request many short ones) – this would make the situation <b>worse</b>, not <b>better</b>.</p> Fri, 08 Aug 2025 08:43:28 +0000 Practical use of LLMs https://lwn.net/Articles/1032930/ https://lwn.net/Articles/1032930/ khim <font class="QuotedText">&gt; In 5-10 years, there will be little distinction between AI assistants and coworkers</font> <p>No. There <b>huge</b> distinction: coworkers can be taught something LLMs would be forever clueless to your objections and requests.</p> <p>You can not teach LLM anything.</p> <p>Well… technically you can, of you train the new model with a [slightly] different properties – but that's not what contributors or coworkers would be able to afford any time soon.</p> <p>And that means that we should insist that contributors (and <b>not</b> maintainers) should shoulder the responsibility of fixing the warts LLMs would add to their submissions again and again.</p> Fri, 08 Aug 2025 08:01:38 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032929/ https://lwn.net/Articles/1032929/ Wol <div class="FormattedComment"> The problem is if the LLM doesn't get it right FIRST TIME, and the person using the LLM doesn't have the experience to recognise this, then it's pretty much assured the result is going to be rubbish. And the chances of this happening with first time submitters is quite high.<br> <p> At the end of the day, most of the stuff on the net is rubbish. The quality of what an LLM outputs is directly correlated to the quality that goes in (it must be, without human review and feedback, it has no clue). Therefor, most LLM output has to be rubbish, too.<br> <p> If your AI is based on a SMALL Language Model, where the stuff fed in has been checked for accuracy, then the results should be pretty decent. I don't use AI at all (as far as I know, the AI search engine slop generally has me going "what aren't you thinking !!!"), but my work now has a little AI that has access to all our help docs and thus does a decent job for most people - except that as always, people don't think, and people keep getting referred to Guru docs for more detail - HINT roughly 1/3 of the company doesn't have access to Guru, as a matter of policy!!! Argh!!!<br> <p> Cheers,<br> Wol<br> </div> Fri, 08 Aug 2025 07:43:07 +0000 Tool use should be tagged in-tree https://lwn.net/Articles/1032928/ https://lwn.net/Articles/1032928/ 奇跡 <blockquote>Jakub Kicinski argued that the information about tools was "only relevant during the review", so putting it into patch changelogs at all "is just free advertising" for the tools in question.</blockquote> This strikes me as an oddly myopic take. Are the drawbacks of such "free advertising" not trivial compared to the obvious auditing/analysis benefits of documenting tool use in tree? Fri, 08 Aug 2025 06:49:02 +0000 Lower bar to start kernel development? https://lwn.net/Articles/1032926/ https://lwn.net/Articles/1032926/ gf2p8affineqb <div class="FormattedComment"> That is quite speculative. The recent rate of growth is no useful indicator for where it will eventually plateau. And right now LLMs are definitely subpar at many programming tasks (also other tasks).<br> </div> Fri, 08 Aug 2025 06:24:05 +0000