<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: Quotations</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/atom/quotations/" rel="self"/><id>http://simonwillison.net/</id><updated>2026-04-28T22:02:53+00:00</updated><author><name>Simon Willison</name></author><entry><title>Quoting OpenAI Codex base_instructions</title><link href="https://simonwillison.net/2026/Apr/28/openai-codex/#atom-quotations" rel="alternate"/><published>2026-04-28T22:02:53+00:00</published><updated>2026-04-28T22:02:53+00:00</updated><id>https://simonwillison.net/2026/Apr/28/openai-codex/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://github.com/openai/codex/blob/66b0781502be5de3b1909525c987643b9e5e407d/codex-rs/models-manager/models.json#L55"&gt;&lt;p&gt;&lt;code&gt;Never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user's query.&lt;/code&gt;&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://github.com/openai/codex/blob/66b0781502be5de3b1909525c987643b9e5e407d/codex-rs/models-manager/models.json#L55"&gt;OpenAI Codex base_instructions&lt;/a&gt;, for GPT-5.5&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/system-prompts"&gt;system-prompts&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-engineering"&gt;prompt-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/codex-cli"&gt;codex-cli&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gpt"&gt;gpt&lt;/a&gt;&lt;/p&gt;

</summary><category term="openai"/><category term="ai"/><category term="llms"/><category term="system-prompts"/><category term="prompt-engineering"/><category term="codex-cli"/><category term="generative-ai"/><category term="gpt"/></entry><entry><title>Quoting Matthew Yglesias</title><link href="https://simonwillison.net/2026/Apr/28/matthew-yglesias/#atom-quotations" rel="alternate"/><published>2026-04-28T13:25:29+00:00</published><updated>2026-04-28T13:25:29+00:00</updated><id>https://simonwillison.net/2026/Apr/28/matthew-yglesias/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://twitter.com/mattyglesias/status/2049105745132585161"&gt;&lt;p&gt;Five months in, I think I've decided that I don't want to vibecode — I want professionally managed software companies to use AI coding assistance to make more/better/cheaper software products that they sell to me for money.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/mattyglesias/status/2049105745132585161"&gt;Matthew Yglesias&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/agentic-engineering"&gt;agentic-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/vibe-coding"&gt;vibe-coding&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;

</summary><category term="agentic-engineering"/><category term="vibe-coding"/><category term="ai-assisted-programming"/><category term="ai"/></entry><entry><title>Quoting Romain Huet</title><link href="https://simonwillison.net/2026/Apr/25/romain-huet/#atom-quotations" rel="alternate"/><published>2026-04-25T12:06:55+00:00</published><updated>2026-04-25T12:06:55+00:00</updated><id>https://simonwillison.net/2026/Apr/25/romain-huet/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://twitter.com/romainhuet/status/2047955381578838357"&gt;&lt;p&gt;Since GPT-5.4, we’ve unified Codex and the main model into a single system, so there’s no separate coding line anymore.&lt;/p&gt;
&lt;p&gt;GPT-5.5 takes this further, with strong gains in agentic coding, computer use, and any task on a computer.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/romainhuet/status/2047955381578838357"&gt;Romain Huet&lt;/a&gt;, confirming OpenAI won't release a GPT-5.5-Codex model&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gpt"&gt;gpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;

</summary><category term="generative-ai"/><category term="gpt"/><category term="openai"/><category term="ai"/><category term="llms"/></entry><entry><title>Quoting Maggie Appleton</title><link href="https://simonwillison.net/2026/Apr/23/maggie-appleton/#atom-quotations" rel="alternate"/><published>2026-04-23T13:35:37+00:00</published><updated>2026-04-23T13:35:37+00:00</updated><id>https://simonwillison.net/2026/Apr/23/maggie-appleton/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://maggieappleton.com/gathering-structures"&gt;&lt;p&gt;[...] if you ever needed another reason to &lt;a href="https://www.swyx.io/learn-in-public"&gt;learn in public&lt;/a&gt; by &lt;a href="https://maggieappleton.com/garden-history"&gt;digital gardening&lt;/a&gt; or podcasting or streaming or whathaveyou, add on that people will assume you’re more competent than you are. This will get you invites to very cool exclusive events filled with high-achieving, interesting people, even though you have no right to be there. A+ side benefit.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://maggieappleton.com/gathering-structures"&gt;Maggie Appleton&lt;/a&gt;, Gathering Structures (&lt;a href="https://notes.andymatuschak.org/Work_with_the_garage_door_up"&gt;via&lt;/a&gt;)&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/blogging"&gt;blogging&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/maggie-appleton"&gt;maggie-appleton&lt;/a&gt;&lt;/p&gt;

</summary><category term="blogging"/><category term="maggie-appleton"/></entry><entry><title>Quoting Bobby Holley</title><link href="https://simonwillison.net/2026/Apr/22/bobby-holley/#atom-quotations" rel="alternate"/><published>2026-04-22T05:40:56+00:00</published><updated>2026-04-22T05:40:56+00:00</updated><id>https://simonwillison.net/2026/Apr/22/bobby-holley/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://blog.mozilla.org/en/privacy-security/ai-security-zero-day-vulnerabilities/"&gt;&lt;p&gt;As part of our continued collaboration with Anthropic, we had the opportunity to apply an early version of Claude Mythos Preview to Firefox. This week’s release of Firefox 150 includes fixes for &lt;a href="https://www.mozilla.org/en-US/security/advisories/mfsa2026-30/"&gt;271 vulnerabilities&lt;/a&gt; identified during this initial evaluation. [...]&lt;/p&gt;
&lt;p&gt;Our experience is a hopeful one for teams who shake off the vertigo and get to work. You may need to reprioritize everything else to bring relentless and single-minded focus to the task, but there is light at the end of the tunnel. We are extremely proud of how our team rose to meet this challenge, and others will too. Our work isn’t finished, but we’ve turned the corner and can glimpse a future much better than just keeping up. &lt;strong&gt;Defenders finally have a chance to win, decisively&lt;/strong&gt;.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://blog.mozilla.org/en/privacy-security/ai-security-zero-day-vulnerabilities/"&gt;Bobby Holley&lt;/a&gt;, CTO, Firefox&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/claude"&gt;claude&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/firefox"&gt;firefox&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/mozilla"&gt;mozilla&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;&lt;/p&gt;

</summary><category term="anthropic"/><category term="claude"/><category term="ai"/><category term="firefox"/><category term="llms"/><category term="mozilla"/><category term="security"/><category term="generative-ai"/><category term="ai-security-research"/></entry><entry><title>Quoting Andreas Påhlsson-Notini</title><link href="https://simonwillison.net/2026/Apr/21/andreas-pahlsson-notini/#atom-quotations" rel="alternate"/><published>2026-04-21T16:39:33+00:00</published><updated>2026-04-21T16:39:33+00:00</updated><id>https://simonwillison.net/2026/Apr/21/andreas-pahlsson-notini/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://nial.se/blog/less-human-ai-agents-please/"&gt;&lt;p&gt;AI agents are already too human. Not in the romantic sense, not because they love or fear or dream, but in the more banal and frustrating one. The current implementations keep showing their human origin again and again: lack of stringency, lack of patience, lack of focus. Faced with an awkward task, they drift towards the familiar. Faced with hard constraints, they start negotiating with reality.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://nial.se/blog/less-human-ai-agents-please/"&gt;Andreas Påhlsson-Notini&lt;/a&gt;, Less human AI agents, please.&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-agents"&gt;ai-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;

</summary><category term="ai-agents"/><category term="coding-agents"/><category term="ai"/></entry><entry><title>Quoting John Gruber</title><link href="https://simonwillison.net/2026/Apr/15/john-gruber/#atom-quotations" rel="alternate"/><published>2026-04-15T17:13:57+00:00</published><updated>2026-04-15T17:13:57+00:00</updated><id>https://simonwillison.net/2026/Apr/15/john-gruber/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://daringfireball.net/2026/04/piece_android_iphone_apps"&gt;&lt;p&gt;The real goldmine isn’t that Apple gets a cut of every App Store transaction. It’s that Apple’s platforms have the best apps, and users who are drawn to the best apps are thus drawn to the iPhone, Mac, and iPad. That edge is waning. Not because software on other platforms is getting better, but because third-party software on iPhone, Mac, and iPad is regressing to the mean, &lt;em&gt;to some extent&lt;/em&gt;, because fewer developers feel motivated — artistically, financially, or both — to create well-crafted idiomatic native apps exclusively for Apple’s platforms.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://daringfireball.net/2026/04/piece_android_iphone_apps"&gt;John Gruber&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/apple"&gt;apple&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/john-gruber"&gt;john-gruber&lt;/a&gt;&lt;/p&gt;

</summary><category term="apple"/><category term="john-gruber"/></entry><entry><title>Quoting Kyle Kingsbury</title><link href="https://simonwillison.net/2026/Apr/15/kyle-kingsbury/#atom-quotations" rel="alternate"/><published>2026-04-15T15:36:02+00:00</published><updated>2026-04-15T15:36:02+00:00</updated><id>https://simonwillison.net/2026/Apr/15/kyle-kingsbury/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://aphyr.com/posts/419-the-future-of-everything-is-lies-i-guess-new-jobs"&gt;&lt;p&gt;I think we will see some people employed (though perhaps not explicitly) as &lt;em&gt;meat shields&lt;/em&gt;: people who are accountable for ML systems under their supervision. The accountability may be purely internal, as when Meta hires human beings to review the decisions of automated moderation systems. It may be external, as when lawyers are penalized for submitting LLM lies to the court. It may involve formalized responsibility, like a Data Protection Officer. It may be convenient for a company to have third-party subcontractors, like Buscaglia, who can be thrown under the bus when the system as a whole misbehaves.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://aphyr.com/posts/419-the-future-of-everything-is-lies-i-guess-new-jobs"&gt;Kyle Kingsbury&lt;/a&gt;, The Future of Everything is Lies, I Guess: New Jobs&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/kyle-kingsbury"&gt;kyle-kingsbury&lt;/a&gt;&lt;/p&gt;

</summary><category term="ai-ethics"/><category term="careers"/><category term="ai"/><category term="kyle-kingsbury"/></entry><entry><title>Quoting Bryan Cantrill</title><link href="https://simonwillison.net/2026/Apr/13/bryan-cantrill/#atom-quotations" rel="alternate"/><published>2026-04-13T02:44:24+00:00</published><updated>2026-04-13T02:44:24+00:00</updated><id>https://simonwillison.net/2026/Apr/13/bryan-cantrill/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://bcantrill.dtrace.org/2026/04/12/the-peril-of-laziness-lost/"&gt;&lt;p&gt;The problem is that LLMs inherently &lt;strong&gt;lack the virtue of laziness&lt;/strong&gt;. Work costs nothing to an LLM. LLMs do not feel a need to optimize for their own (or anyone's) future time, and will happily dump more and more onto a layercake of garbage. Left unchecked, LLMs will make systems larger, not better &amp;mdash; appealing to perverse vanity metrics, perhaps, but at the cost of everything that matters.&lt;/p&gt;
&lt;p&gt;As such, LLMs highlight how essential our human laziness is: our finite time &lt;strong&gt;forces&lt;/strong&gt; us to develop crisp abstractions in part because we don't want to waste our (human!) time on the consequences of clunky ones.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://bcantrill.dtrace.org/2026/04/12/the-peril-of-laziness-lost/"&gt;Bryan Cantrill&lt;/a&gt;, The peril of laziness lost&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bryan-cantrill"&gt;bryan-cantrill&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;&lt;/p&gt;

</summary><category term="bryan-cantrill"/><category term="ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="generative-ai"/></entry><entry><title>Quoting Giles Turnbull</title><link href="https://simonwillison.net/2026/Apr/8/giles-turnbull/#atom-quotations" rel="alternate"/><published>2026-04-08T15:18:49+00:00</published><updated>2026-04-08T15:18:49+00:00</updated><id>https://simonwillison.net/2026/Apr/8/giles-turnbull/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://gilest.org/notes/2026/human-ai/"&gt;&lt;p&gt;I have a feeling that &lt;strong&gt;everyone likes using AI tools to try doing someone else’s profession&lt;/strong&gt;. They’re much less keen when someone else uses it for their profession.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://gilest.org/notes/2026/human-ai/"&gt;Giles Turnbull&lt;/a&gt;, AI and the human voice&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/writing"&gt;writing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;

</summary><category term="ai-ethics"/><category term="writing"/><category term="ai"/></entry><entry><title>Quoting Chengpeng Mou</title><link href="https://simonwillison.net/2026/Apr/5/chengpeng-mou/#atom-quotations" rel="alternate"/><published>2026-04-05T21:47:06+00:00</published><updated>2026-04-05T21:47:06+00:00</updated><id>https://simonwillison.net/2026/Apr/5/chengpeng-mou/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://twitter.com/cpmou2022/status/2040606209800290404"&gt;&lt;p&gt;From anonymized U.S. ChatGPT data, we are seeing:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;~2M weekly messages on health insurance&lt;/li&gt;
&lt;li&gt;~600K weekly messages [classified as healthcare] from people living in “hospital deserts” (30 min drive to nearest hospital)&lt;/li&gt;
&lt;li&gt;7 out of 10 msgs happen outside clinic hours&lt;/li&gt;
&lt;/ul&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/cpmou2022/status/2040606209800290404"&gt;Chengpeng Mou&lt;/a&gt;, Head of Business Finance, OpenAI&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;

</summary><category term="ai-ethics"/><category term="generative-ai"/><category term="openai"/><category term="chatgpt"/><category term="ai"/><category term="llms"/></entry><entry><title>Quoting Kyle Daigle</title><link href="https://simonwillison.net/2026/Apr/4/kyle-daigle/#atom-quotations" rel="alternate"/><published>2026-04-04T02:20:17+00:00</published><updated>2026-04-04T02:20:17+00:00</updated><id>https://simonwillison.net/2026/Apr/4/kyle-daigle/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://twitter.com/kdaigle/status/2040164759836778878"&gt;&lt;p&gt;[GitHub] platform activity is surging. There were 1 billion commits in 2025. Now, it's 275 million per week, on pace for 14 billion this year if growth remains linear (spoiler: it won't.)&lt;/p&gt;
&lt;p&gt;GitHub Actions has grown from 500M minutes/week in 2023 to 1B minutes/week in 2025, and now 2.1B minutes so far this week.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/kdaigle/status/2040164759836778878"&gt;Kyle Daigle&lt;/a&gt;, COO, GitHub&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/github"&gt;github&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/github-actions"&gt;github-actions&lt;/a&gt;&lt;/p&gt;

</summary><category term="github"/><category term="github-actions"/></entry><entry><title>Quoting Willy Tarreau</title><link href="https://simonwillison.net/2026/Apr/3/willy-tarreau/#atom-quotations" rel="alternate"/><published>2026-04-03T21:48:22+00:00</published><updated>2026-04-03T21:48:22+00:00</updated><id>https://simonwillison.net/2026/Apr/3/willy-tarreau/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://lwn.net/Articles/1065620/"&gt;&lt;p&gt;On the kernel security list we've seen a huge bump of reports. We were between 2 and 3 per week maybe two years ago, then reached probably 10 a week over the last year with the only difference being only AI slop, and now since the beginning of the year we're around 5-10 per day depending on the days (fridays and tuesdays seem the worst). Now most of these reports are correct, to the point that we had to bring in more maintainers to help us.&lt;/p&gt;
&lt;p&gt;And we're now seeing on a daily basis something that never happened before: duplicate reports, or the same bug found by two different people using (possibly slightly) different tools.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://lwn.net/Articles/1065620/"&gt;Willy Tarreau&lt;/a&gt;, Lead Software Developer. HAPROXY&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/linux"&gt;linux&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;&lt;/p&gt;

</summary><category term="security"/><category term="linux"/><category term="generative-ai"/><category term="ai"/><category term="llms"/><category term="ai-security-research"/></entry><entry><title>Quoting Daniel Stenberg</title><link href="https://simonwillison.net/2026/Apr/3/daniel-stenberg/#atom-quotations" rel="alternate"/><published>2026-04-03T21:46:07+00:00</published><updated>2026-04-03T21:46:07+00:00</updated><id>https://simonwillison.net/2026/Apr/3/daniel-stenberg/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://mastodon.social/@bagder/116336957584445742"&gt;&lt;p&gt;The challenge with AI in open source security has transitioned from an AI slop tsunami into more of a ... plain security report tsunami. Less slop but lots of reports. Many of them really good.&lt;/p&gt;
&lt;p&gt;I'm spending hours per day on this now. It's intense.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://mastodon.social/@bagder/116336957584445742"&gt;Daniel Stenberg&lt;/a&gt;, lead developer of cURL&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/daniel-stenberg"&gt;daniel-stenberg&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/curl"&gt;curl&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;&lt;/p&gt;

</summary><category term="daniel-stenberg"/><category term="security"/><category term="curl"/><category term="generative-ai"/><category term="ai"/><category term="llms"/><category term="ai-security-research"/></entry><entry><title>Quoting Greg Kroah-Hartman</title><link href="https://simonwillison.net/2026/Apr/3/greg-kroah-hartman/#atom-quotations" rel="alternate"/><published>2026-04-03T21:44:41+00:00</published><updated>2026-04-03T21:44:41+00:00</updated><id>https://simonwillison.net/2026/Apr/3/greg-kroah-hartman/#atom-quotations</id><summary type="html">&lt;blockquote cite="https://www.theregister.com/2026/03/26/greg_kroahhartman_ai_kernel/"&gt;&lt;p&gt;Months ago, we were getting what we called 'AI slop,' AI-generated security reports that were obviously wrong or low quality. It was kind of funny. It didn't really worry us.&lt;/p&gt;
&lt;p&gt;Something happened a month ago, and the world switched. Now we have real reports. All open source projects have real reports that are made with AI, but they're good, and they're real.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://www.theregister.com/2026/03/26/greg_kroahhartman_ai_kernel/"&gt;Greg Kroah-Hartman&lt;/a&gt;, Linux kernel maintainer (&lt;a href="https://en.wikipedia.org/wiki/Greg_Kroah-Hartman"&gt;bio&lt;/a&gt;), in conversation with Steven J. Vaughan-Nichols&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/linux"&gt;linux&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;&lt;/p&gt;

</summary><category term="security"/><category term="linux"/><category term="generative-ai"/><category term="ai"/><category term="llms"/><category term="ai-security-research"/></entry></feed>