<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: propublica</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/propublica.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2026-02-13T23:38:29+00:00</updated><author><name>Simon Willison</name></author><entry><title>The evolution of OpenAI's mission statement</title><link href="https://simonwillison.net/2026/Feb/13/openai-mission-statement/#atom-tag" rel="alternate"/><published>2026-02-13T23:38:29+00:00</published><updated>2026-02-13T23:38:29+00:00</updated><id>https://simonwillison.net/2026/Feb/13/openai-mission-statement/#atom-tag</id><summary type="html">
    &lt;p&gt;As a USA &lt;a href="https://en.wikipedia.org/wiki/501(c)(3)_organization"&gt;501(c)(3)&lt;/a&gt; the OpenAI non-profit has to file a tax return each year with the IRS. One of the required fields on that tax return is to "Briefly describe the organization’s mission or most significant activities" - this has actual legal weight to it as the IRS can use it to evaluate if the organization is sticking to its mission and deserves to maintain its non-profit tax-exempt status.&lt;/p&gt;
&lt;p&gt;You can browse OpenAI's &lt;a href="https://projects.propublica.org/nonprofits/organizations/810861541"&gt;tax filings by year&lt;/a&gt; on ProPublica's excellent &lt;a href="https://projects.propublica.org/nonprofits/"&gt;Nonprofit Explorer&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;I went through and extracted that mission statement for 2016 through 2024, then had Claude Code &lt;a href="https://gisthost.github.io/?7a569df89f43f390bccc2c5517718b49/index.html"&gt;help me&lt;/a&gt; fake the commit dates to turn it into a git repository and share that as a Gist - which means that Gist's &lt;a href="https://gist.github.com/simonw/e36f0e5ef4a86881d145083f759bcf25/revisions"&gt;revisions page&lt;/a&gt; shows every edit they've made since they started filing their taxes!&lt;/p&gt;
&lt;p&gt;It's really interesting seeing what they've changed over time.&lt;/p&gt;
&lt;p&gt;The original 2016 mission reads as follows (and yes, the apostrophe in "OpenAIs" is missing &lt;a href="https://projects.propublica.org/nonprofits/organizations/810861541/201703459349300445/full"&gt;in the original&lt;/a&gt;):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;OpenAIs goal is to advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return. We think that artificial intelligence technology will help shape the 21st century, and we want to help the world build safe AI technology and ensure that AI's benefits are as widely and evenly distributed as possible. Were trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;In 2018 they dropped the part about "trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way."&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-3.jpg" alt="Git diff showing the 2018 revision deleting the final two sentences: &amp;quot;Were trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way.&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;In 2020 they dropped the words "as a whole" from "benefit humanity as a whole". They're still "unconstrained by a need to generate financial return" though.&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-5.jpg" alt="Git diff showing the 2020 revision dropping &amp;quot;as a whole&amp;quot; from &amp;quot;benefit humanity as a whole&amp;quot; and changing &amp;quot;We think&amp;quot; to &amp;quot;OpenAI believes&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;Some interesting changes in 2021. They're still unconstrained by a need to generate financial return, but here we have the first reference to "general-purpose artificial intelligence" (replacing "digital intelligence"). They're more confident too: it's not "most likely to benefit humanity", it's just "benefits humanity".&lt;/p&gt;
&lt;p&gt;They previously wanted to "help the world build safe AI technology", but now they're going to do that themselves: "the companys goal is to develop and responsibly deploy safe AI technology".&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-6.jpg" alt="Git diff showing the 2021 revision replacing &amp;quot;goal is to advance digital intelligence&amp;quot; with &amp;quot;mission is to build general-purpose artificial intelligence&amp;quot;, changing &amp;quot;most likely to benefit&amp;quot; to just &amp;quot;benefits&amp;quot;, and replacing &amp;quot;help the world build safe AI technology&amp;quot; with &amp;quot;the companys goal is to develop and responsibly deploy safe AI technology&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;2022 only changed one significant word: they added "safely" to "build ... (AI) that safely benefits humanity". They're still unconstrained by those financial returns!&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-7.jpg" alt="Git diff showing the 2022 revision adding &amp;quot;(AI)&amp;quot; and the word &amp;quot;safely&amp;quot; so it now reads &amp;quot;that safely benefits humanity&amp;quot;, and changing &amp;quot;the companys&amp;quot; to &amp;quot;our&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;No changes in 2023... but then in 2024 they deleted almost the entire thing, reducing it to simply:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;OpenAIs mission is to ensure that artificial general intelligence benefits all of humanity.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;They've expanded "humanity" to "all of humanity", but there's no mention of safety any more and I guess they can finally start focusing on that need to generate financial returns!&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-9.jpg" alt="Git diff showing the 2024 revision deleting the entire multi-sentence mission statement and replacing it with just &amp;quot;OpenAIs mission is to ensure that artificial general intelligence benefits all of humanity.&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Update&lt;/strong&gt;: I found loosely equivalent but much less interesting documents &lt;a href="https://simonwillison.net/2026/Feb/13/anthropic-public-benefit-mission/"&gt;from Anthropic&lt;/a&gt;.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/propublica"&gt;propublica&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="ai"/><category term="openai"/><category term="ai-ethics"/><category term="propublica"/></entry><entry><title>How ProPublica Uses AI Responsibly in Its Investigations</title><link href="https://simonwillison.net/2025/Mar/14/propublica-ai/#atom-tag" rel="alternate"/><published>2025-03-14T21:04:46+00:00</published><updated>2025-03-14T21:04:46+00:00</updated><id>https://simonwillison.net/2025/Mar/14/propublica-ai/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.propublica.org/article/using-ai-responsibly-for-reporting"&gt;How ProPublica Uses AI Responsibly in Its Investigations&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Charles Ornstein describes how ProPublica used an LLM to help analyze data for their recent story &lt;a href="https://www.propublica.org/article/ted-cruz-woke-grants-national-science-foundation"&gt;A Study of Mint Plants. A Device to Stop Bleeding. This Is the Scientific Research Ted Cruz Calls “Woke.”&lt;/a&gt; by Agnel Philip and Lisa Song.&lt;/p&gt;
&lt;p&gt;They ran ~3,400 grant descriptions through a prompt that included the following:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;As an investigative journalist, I am looking for the following information&lt;/p&gt;
&lt;p&gt;-- &lt;/p&gt;
&lt;p&gt;&lt;code&gt;woke_description&lt;/code&gt;: A short description (at maximum a paragraph) on why this grant is being singled out for promoting "woke" ideology, Diversity, Equity, and Inclusion (DEI) or advanced neo-Marxist class warfare propaganda. Leave this blank if it's unclear.&lt;/p&gt;
&lt;p&gt;&lt;code&gt;why_flagged&lt;/code&gt;: Look at the "STATUS", "SOCIAL JUSTICE CATEGORY", "RACE CATEGORY", "GENDER CATEGORY" and "ENVIRONMENTAL JUSTICE CATEGORY" fields. If it's filled out, it means that the author of this document believed the grant was promoting DEI ideology in that way. Analyze the "AWARD DESCRIPTIONS" field and see if you can figure out why the author may have flagged it in this way. Write it in a way that is thorough and easy to understand with only one description per type and award.&lt;/p&gt;
&lt;p&gt;&lt;code&gt;citation_for_flag&lt;/code&gt;: Extract a very concise text quoting the passage of "AWARDS DESCRIPTIONS" that backs up the "why_flagged" data.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;This was only the first step in the analysis of the data:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Of course, members of our staff reviewed and confirmed every detail before we published our story, and we called all the named people and agencies seeking comment, which remains a must-do even in the world of AI.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I think journalists are particularly well positioned to take advantage of LLMs in this way, because a big part of journalism is about deriving the truth from multiple unreliable sources of information. Journalists are deeply familiar with fact-checking, which is a critical skill if you're going to report with the assistance of these powerful but unreliable models.&lt;/p&gt;
&lt;p&gt;Agnel Philip:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The tech holds a ton of promise in lead generation and pointing us in the right direction. But in my experience, it still needs a lot of human supervision and vetting. If used correctly, it can both really speed up the process of understanding large sets of information, and if you’re creative with your prompts and critically read the output, it can help uncover things that you may not have thought of.&lt;/p&gt;
&lt;/blockquote&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/data-journalism"&gt;data-journalism&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ethics"&gt;ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/journalism"&gt;journalism&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-engineering"&gt;prompt-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/structured-extraction"&gt;structured-extraction&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/propublica"&gt;propublica&lt;/a&gt;&lt;/p&gt;



</summary><category term="data-journalism"/><category term="ethics"/><category term="journalism"/><category term="ai"/><category term="prompt-engineering"/><category term="generative-ai"/><category term="llms"/><category term="structured-extraction"/><category term="ai-ethics"/><category term="propublica"/></entry></feed>