<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Can We Secure AI With Formal Methods?]]></title><description><![CDATA[Formal methods needs to know that AI security folks are a critical fountain of users. AI security folks need to know how to ask formal methodsititians for widgets. FKA Progress in Guaranteed Safe AI.
]]></description><link>https://newsletter.for-all.dev</link><image><url>https://substackcdn.com/image/fetch/$s_!ykg_!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F083283b3-e660-4c7f-81e3-c40b1d1ebafd_1024x1024.png</url><title>Can We Secure AI With Formal Methods?</title><link>https://newsletter.for-all.dev</link></image><generator>Substack</generator><lastBuildDate>Sun, 12 Apr 2026 04:40:10 GMT</lastBuildDate><atom:link href="https://newsletter.for-all.dev/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Quinn Dougherty]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[gsai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[gsai@substack.com]]></itunes:email><itunes:name><![CDATA[Quinn Dougherty]]></itunes:name></itunes:owner><itunes:author><![CDATA[Quinn Dougherty]]></itunes:author><googleplay:owner><![CDATA[gsai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[gsai@substack.com]]></googleplay:email><googleplay:author><![CDATA[Quinn Dougherty]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Can We Secure AI With Formal Methods? January-March 2026]]></title><description><![CDATA[I'm overwhelmed, cut me some slack.]]></description><link>https://newsletter.for-all.dev/p/can-we-secure-ai-with-formal-methods-480</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/can-we-secure-ai-with-formal-methods-480</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Thu, 02 Apr 2026 18:41:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ctuV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the month or so around the previous new years, as 2024 became 2025, we were saying &#8220;2025: year of the agent&#8221;. MCP was taking off, the inspect-ai and pydantic-ai python packages were becoming the standards, products were branching out from chatbots to heavy and autonomous use of toolcalls. While much of the product engineering scene may have underdelivered (in the sense that &#8220;planning a vacation&#8221; isn&#8217;t entirely something most people do with agents yet), the field of FMxAI I think was right on target. Feels like there&#8217;s an agentic component to everything I read these days.</p><p>What is 2026 the year of? Besides &#8220;year of investors pressure all the math companies to pivot to program synthesis&#8221;? I&#8217;m declaring it now</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ctuV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ctuV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 424w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 848w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 1272w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ctuV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png" width="668" height="374" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/05c03da8-f543-4344-8f02-c32710a7d251_668x374.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:374,&quot;width&quot;:668,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ctuV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 424w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 848w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 1272w, https://substackcdn.com/image/fetch/$s_!ctuV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05c03da8-f543-4344-8f02-c32710a7d251_668x374.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The number of blogposts relating to secure program synthesis <a href="https://www.lesswrong.com/posts/SfhFh9Hfm6JYvzbby/the-scalable-formal-oversight-research-program">went</a> <a href="https://john.regehr.org/writing/zero_dof_programming.html">exponential</a> <a href="https://martin.kleppmann.com/2025/12/08/ai-formal-verification.html">since</a> <a href="https://leodemoura.github.io/blog/2026-3-16-who-watches-the-provers/">the</a> <a href="https://www.lesswrong.com/posts/8wtrLoDPyCfMLuHkt/how-to-solve-secure-program-synthesis">last</a> <a href="https://x.com/karpathy/status/2023476423055601903">issue</a> <a href="https://www.amplifypartners.com/blog-posts/the-agentic-mullet-code-in-the-front-proofs-in-the-back">of</a> the newsletter, such that the hyperlinks in this sentence are not even exhaustive. Like AI capabilities themselves, we simply do not know when it will go logistic. Which I think means I have to keep the newsletter manageable by not promising to do the general secure program synthesis or FMxAI spaces and really double down on AI security applications. I mean it, you guys. It&#8217;s overwhelming, the space is too big now, etc. But <strong>will SPS gains actually get applied to AI security opportunities?</strong> That&#8217;s the crucial question I think we should care about, now. Let&#8217;s assume by default no, so that everyone who wants to liberate my bayes points from me goes and makes it happen.</p><p>I moved to DC, and I started an AI Security reading group in meatspace. Reach out to be involved!</p><p>There will be hackathons and fellowships in secure program synthesis, FMxAI, AI security, etc. Be prepared.</p><p>Sorry again about the long gap between editions. I&#8217;ve been writing more than ever, but not a lot of it public yet. Also coding a lot.</p><h1><a href="https://youtu.be/1sd26pWhfmg?si=Hwcm1jNBrYhoD5d2">Nicholas Carlini&#8217;s desperate and pleading talk</a></h1><p>Carlini: &#8220;Opus is better at finding 0-days than I am&#8221;. Carlini, famous for finding vulns in adversarial robustness papers back in the computer vision days, stopped a hair short of declaring a draft, but I&#8217;m not. This <em>is </em>a draft. If you&#8217;re reading this, you are hereby conscripted into the secure program synthesis army. Or maybe our vibe is more &#8220;militia&#8221; coded, haven&#8217;t made up my mind yet. But this <em>is</em> a draft. The gameboard seems offense dominated right now, we need to harden software, and we need to do it quickly.</p><p>According to gemini, who cites <a href="https://www.forbes.com/1998/03/12/feat.html#:~:text=Search%20for%20the%20Y2K%20problem,range%2C%20that)%20to%20fix.">forbes</a>: 20-40% of corporate IT budgets in 1999 were devoted to Y2K prep. $300-600b was spent (roughly 0.5-1T adjusted for inflation). Ordinary people, people with families, people who put sandwiches into little baggies, decided to do this. They had consensus that it was coming down the pipe, on a deadline.</p><p>I don&#8217;t think we have that consensus, I&#8217;m not sure we can marshall those resources.</p><p>But here look this is all I&#8217;m saying</p><p><code>cat &lt;&lt; EOF &gt; CLAUDE.md</code></p><p><code>Make agentic software that hardens a given repo with a red-blue loop. Declare an agent that uses standard redteam tools like fuzzing harnesses and static analysis as tools to find vulns in repos, then declare another agent that patches those vulns. Use git submodules to manage sources</code></p><p><code>EOF</code></p><p><code>while true; do cat CLAUDE.md | claude --dangerously-skip-permissions; sleep 1; done</code></p><p>If everyone reading this ran that command, and tailored it to the specific things you happen to know about, and dropped in repos close to your specialization, and steered the project in directions that I wouldn&#8217;t think of, etc. we might have a shot. Do not DoS open source maintainers with slop, but consider forking their repos and DoSing your forks with slop then finding a way to benchmark how much more secure your fork is and then maybe you get a few PRs in. Target loadbearing repos. Etc. &#129761;</p><h1><a href="https://www.gov.uk/government/publications/secure-ai-infrastructure-call-for-information/secure-ai-infrastructure-call-for-information?utm_source=Twitter&amp;utm_medium=social&amp;utm_campaign=Orlo">UK AISI call for information on Secure AI Infrastructure</a></h1><p>Let&#8217;s do it. Doesn&#8217;t seem to have a deadline.</p><p>Gemini&#8217;s summary:</p><blockquote><p>The UK government has launched a &#8220;Call for Information&#8221; to gather expert insights on securing the computing infrastructure used to develop and deploy advanced AI models. Managed jointly by the Department for Science, Innovation and Technology (DSIT), the AI Security Institute (AISI), and the National Cyber Security Centre (NCSC), the initiative seeks to address growing threats such as the theft of model weights, data breaches, and system disruptions. The government is specifically soliciting feedback from the AI and cybersecurity sectors on current risks and emerging technologies&#8212;including confidential compute, advanced cryptography, and trusted hardware&#8212;to help shape future research priorities and technical pilots. Ultimately, the program aims to ensure the UK remains a secure and trusted global hub for frontier AI development by building robust, &#8220;defense-in-depth&#8221; protections for critical AI assets.</p></blockquote><h1><a href="https://sl5.org/projects">SL5 Taskforce has been productive!</a></h1><p>Excited about some of these outputs.</p><h1>Long Live Safeguarded AI</h1><p><a href="https://aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai/">Official leadership update</a>, <a href="https://vimeo.com/1178834361">Kathleen Fisher (ex DARPA, responsible for HACMS) interviews Nora and Davidad on vimeo</a>, <a href="https://x.com/davidad/status/2039390998694891816">Davidad&#8217;s notes about what he&#8217;s doing next</a>.</p><p>Safeguarded AI, like GSAI, like Open Agency Architecture, was always controversial scifi that, to be clear, I&#8217;ve been rooting for since day one. Well not literally day one. When I first read open agency architecture, I recall rolling my eyes and wringing my hands &#8220;you can&#8217;t just put a type signature on a wacky diagram and say that it means AI safety is formally verifiable&#8221; but like a few days later I was into it.</p><p>With Nora Ammann in as programme director (who&#8217;s been behind the scenes on SGAI since before day one), we should expect aggressive AI security applications of formal methods.</p><h1><a href="https://www.darpa.mil/about/people/patrick-lincoln">Pat Lincoln I2O director at DARPA</a></h1><p>Totally should&#8217;ve been in the thanksgiving edition, I spaced. <a href="https://www.sri.com/">SRI</a> alum Pat Lincoln is pretty keen about AGI, he&#8217;s not like completely AGI pilled but extremely sympathetic to the viewpoint.</p><h1><a href="https://blog.atlascomputing.org/p/cslib-leans-formal-software-foundation">Atlas blog on CSLib</a></h1><p><a href="https://www.cslib.io/">CSLib</a> is what it sounds like: the mathlib of CS. It is spearheaded by Clark Barrett and Swarat Chaudhuri and others. Atlas Computing, which has <a href="https://newsletter.for-all.dev/i/148170804/atlas-ships-their-big-google-doc-alluded-to-in-the-last-newsletter">for a while</a> been looking at how to close the gap between formal methods as we find them today and radical infrastructure hardening, writes about it (they&#8217;re directly involved).</p><p>CSLib is cool, but one reason I think it doesn&#8217;t register as the droid I&#8217;m looking for is that I don&#8217;t know if real world software really draws from undergrad CS curriculum all that much. In most real world software jobs, having a problem that is shaped like an academic puzzle rather than a REST API is a rare but golden status marker.</p><h1><a href="https://utopia-group.github.io/VeriSoftBench/">VeriSoftBench</a></h1><p>Claim: idiosyncrasies of mathlib in particular determine too much of what we think we know about Lean evals and language model performance on them.</p><blockquote><p>500 theorem-proving tasks from 23 real-world projects, spanning compilers, type systems, smart contracts, separation logic, and program semantics.</p></blockquote><p>Recognizing the crucial wisdom of our time (that math is made of tokens which we can use for something else), we get a Lean eval that isn&#8217;t focused on math.</p><h1><a href="https://medium.com/@rahatesamruddhi7/%CE%BCinference-building-an-sl5-weight-enclave-on-sel4-5d044f832ee6">muInference: a minimal inference stack on SeL4</a></h1><p>In the draft of this newsletter, where I stubbed out writing about this as TODO, the stub consisted of one word:</p><blockquote><p>Dude.</p></blockquote><p>Which I think roughly describes how I felt seeing it and how I feel now. They basically boot a minimal (cpu) inference enclave in seL4. I used it as an enclave in a network simulation project I was doing, and it doesn&#8217;t even have a machine-readable TTY, <em>that</em>&#8217;s how secure it is.</p><h1>Progress in the mitigation of steganography</h1><p>Crucial applications for model weight exfiltration defense!</p><h2><a href="https://arxiv.org/abs/2511.20621">DiFR: Inference Verification Despite Nondeterminism</a></h2><h2><a href="https://arxiv.org/abs/2511.02620">Verifying LLM Inference to Detect Model Weight Exfiltration</a></h2><h1><a href="https://seldonlab.com/">Seldon Labs batch 2 applications</a> (closed by the time I&#8217;m posting this)</h1><p>Sorry I didn&#8217;t get this out fast enough. But keep an eye on Seldon re the general space of AI security startups.</p><h1>Also keep an eye on: <a href="https://www.heronsec.ai">Heron</a></h1><p>Met a bunch of these fine folks at EAG.</p><h1><a href="https://lucidcomputing.ai/labs">Lucid Labs: data center access program for hardware related experiments</a></h1><p>Summary by gemini:</p><blockquote><p>The <strong>Lab Access Program</strong> by Lucid Computing is a specialized initiative providing developers and enterprises with early, hands-on access to &#8220;sovereign&#8221; AI infrastructure built on zero-trust, hardware-rooted security. Designed for industries with strict regulatory requirements&#8212;such as defense, healthcare, and finance&#8212;the program offers a sandbox environment where users can deploy AI models and agents while maintaining cryptographic proof of compliance and data residency. By moving away from traditional &#8220;black box&#8221; cloud models, participants can verify that their computations are occurring on specific, untampered hardware, allowing them to benchmark and refine secure AI workflows before moving into full-scale production.</p></blockquote><h1><a href="https://www.lesswrong.com/posts/7aJwgbMEiKq5egQbd/ai-found-12-of-12-openssl-zero-days-while-curl-cancelled-its">AISLE finds 12 0-days in openssl and doesn&#8217;t (appear to) use formal methods</a></h1><p>The hero the secure program synthesis community has been waiting for.</p><p>Its extremely unlikely that theyre using iris or some synthetic lean-iris under the hood. A year ago at EAG 2025, I went around telling everybody about my budding separation logic agent and that I wanted to find bugs, synthesize patches, and prove the patches correct with it&#8211; and I used repos like OpenSSL as hypothetical target. I didn&#8217;t continue working on that cuz I didnt like the Iris DSL that it was based on, but I&#8217;m not sure if I should&#8217;ve tried to continue.</p><h1>Take a look at <a href="http://aisafety.com/">aisafety.com</a> for overall fieldmapping</h1><p>Thanks to Bryce and the folks at <a href="http://aisafety.com">aisafety.com</a>, a bunch of you have found this newsletter, where we discuss using formal methods to improve our AI security posture. Many of the rest of you, however, might have come in from the formal methods community and sometimes think &#8220;who are all you AI existential safety wackos&#8221;. <a href="http://aisafety.com">aisafety.com</a> wants to answer that question by making a map of everything that&#8217;s going on in the <em>very broad</em> umbrella of AI existential safety. Highly recommended if you&#8217;re generally scared about AI but don&#8217;t know exactly where to plug in yet.</p><h1><a href="https://ifp.org/rfp-launch/">Allies IFP have new RFP</a></h1><p>With a great track record of supporting secure program synthesis, IFP wants to do more projects along the lines of their galvanization of <a href="https://ifp.org/the-great-refactor/">the great refactor</a>.</p><h1><a href="https://www.rand.org/pubs/commentary/2026/02/tipping-the-cyber-balance-how-ai-benchmarks-could-make.html">Kathleen Fisher and Gopal Sarma post op-ed</a></h1><p>Let&#8217;s ship evals and RL environments <em>now</em> so that the next gen of language models are differentially accelerated to cyberhardening defensive technologies. 10/10, no notes.</p>]]></content:encoded></item><item><title><![CDATA[Can We Secure AI With Formal Methods? November-December 2025]]></title><description><![CDATA[Give thanks for benchmarks with &#8220;Veri-&#8221; in the title.]]></description><link>https://newsletter.for-all.dev/p/can-we-secure-ai-with-formal-methods</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/can-we-secure-ai-with-formal-methods</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Sat, 29 Nov 2025 14:04:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!q5Ef!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>We did the rebrand! The previous thumbnail was a baseball metaphor, but it was very clearly someone getting <em>out</em>, not <em>safe</em>. I was testing all of you and each of you FAILED.</p><p>Here&#8217;s the prompt for the new thumbnail:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Can We Secure AI With Formal Methods? is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><blockquote><p>i&#8217;m keeping AI in a box, doing AI CConfinement (like in yampolskiy 2012), using formal verification / formal methods. That&#8217;s my whole thing. I need art for my newsletter on these topics. I like the percival story from troyes/wagner and i like tolkien, but if you take from those elements put it IN SPACE like scifi. Also use german expressionist painting styles. Ok now give me some DALLE art.</p></blockquote><p>So long &#8220;Progress in GSAI&#8221;. I still like the <a href="https://arxiv.org/abs/2405.06624">position paper</a> that the old newsletter title was based on, but</p><ol><li><p>It&#8217;s very scifi and I think there&#8217;s more alpha in obvious/relatively easy/uncontroversial (but not done by default) work.</p></li><li><p>The word &#8220;guarantee&#8221; doesn&#8217;t evoke &#8220;<a href="https://www.lesswrong.com/posts/CCT7Qc8rSeRs7r5GL/beliefs-about-formal-methods-and-ai-safety#Swiss_cheese__">swiss cheese</a>&#8221;.</p></li><li><p>It&#8217;s time to double down on relationships between AI security and formal methods, directly, more explicitly than you can do within the framing of GSAI.</p></li></ol><p>Also notice: <a href="http://gsai.substack.com">gsai.substack.com</a> is now a redirect to newsletter.for-all.dev. I&#8217;ll be hosting a bunch of my technical reports and comms/strategy outputs at that domain going forward (the subdomain newsletter will just point to substack). But don&#8217;t worry, the scope of the newsletter remains largely the same (excepting the pivot to be more directly and explicitly about AI security) / won&#8217;t devolve into being any more nakedly self promotional than it has been so far.</p><p>I received a grant from a Funder of Presently Undisclosed Provenance to do comms and strategy for AI security via formal methods, which means among other things that this newsletter will get a little more TLC.</p><p>Busy month, I expect things to be slow over christmas, after this edition I&#8217;ll see you all in 2026.</p><p>In the spirit of chivalry, I styletransferred most abstracts in this edition of the newsletter to Troyes/Cervantes style. I did not check to see if Gemini got anything wrong, but every headline is a link to arxiv or openreview which you&#8217;ll click if you&#8217;re interested.</p><h1><a href="https://arxiv.org/abs/2511.10783">Miri&#8217;s treaty team posts a paper!</a></h1><p>Excited about this. They use the word &#8220;verification&#8221; in a different context then we do, they mean it in the sense of verifying the absence of enriched uranium (GPUs) or verifying that the terms of a treaty are being abided by.</p><blockquote><p>Many experts argue that premature development of artificial superintelligence (ASI) poses catastrophic risks, including the risk of human extinction from misaligned ASI, geopolitical instability, and misuse by malicious actors. This report proposes an international agreement to prevent the premature development of ASI until AI development can proceed without these risks. The agreement halts dangerous AI capabilities advancement while preserving access to current, safe AI applications.</p><p>The proposed framework centers on a coalition led by the United States and China that would restrict the scale of AI training and dangerous AI research. Due to the lack of trust between parties, verification is a key part of the agreement. Limits on the scale of AI training are operationalized by FLOP thresholds and verified through the tracking of AI chips and verification of chip use. Dangerous AI research--that which advances toward artificial superintelligence or endangers the agreement&#8217;s verifiability--is stopped via legal prohibitions and multifaceted verification.</p><p>We believe the proposal would be technically sufficient to forestall the development of ASI if implemented today, but advancements in AI capabilities or development methods could hurt its efficacy. Additionally, there does not yet exist the political will to put such an agreement in place. Despite these challenges, we hope this agreement can provide direction for AI governance research and policy.</p></blockquote><h1><a href="https://bluerocksec.gitlab.io/formal-methods/">BlueRock GPLs the specs and proofs of NOVA</a></h1><p><a href="https://bluerocksec.gitlab.io/formal-methods/blogs/2025-10-16-NOVA-specs-proofs-open/">Three</a>. <a href="https://bluerocksec.gitlab.io/formal-methods/blogs/2025-11-18-NOVA-proof-spec/">Great</a>. <a href="https://bluerocksec.gitlab.io/formal-methods/blogs/2025-11-18-NOVA-october-release/">Blog posts</a>. The third one of interest for insight into maintenance and repair of a spec and proof codebase.</p><p>NOVA is the legendary <a href="https://hypervisor.org/">hypervisor</a> that was specified and proven correct at BlueRock (FKA Bedrock). I say &#8220;legendary&#8221; because as a wee lad, stalking Bedrock&#8217;s github activity, hearing rumors about C++ verification, it was one of the few Ws of industrial verification at scale that I had heard about.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q5Ef!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q5Ef!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 424w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 848w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 1272w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q5Ef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png" width="860" height="853" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:853,&quot;width&quot;:860,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q5Ef!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 424w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 848w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 1272w, https://substackcdn.com/image/fetch/$s_!q5Ef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2906eac6-843f-40e4-8754-e231c5ce6d3f_860x853.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Look at that <a href="https://gitlab.com/bluerocksec/NOVA/-/blob/proof/ver/spec/rocq-bluerock-lts-adequacy/theories/lts/x86_cpu.v">B-E-A-YOOT</a>.</p><p>A hypervisor is a part of the virtual machine stack. NOVA is a hardened one for critical systems, technically a microhypervisor.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4YX9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4YX9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 424w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 848w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 1272w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4YX9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png" width="621" height="941" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:941,&quot;width&quot;:621,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4YX9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 424w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 848w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 1272w, https://substackcdn.com/image/fetch/$s_!4YX9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb669c5ad-26ff-4b52-8b3f-f228658d6868_621x941.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We should teach AIs to write this stuff, cuz that looks painful to type.</p><p>We don&#8217;t talk enough about separation logic here on the newsletter. Anyways,</p><h1>People are playing with <a href="https://aristotle.harmonic.fun/">Aristotle</a>, Harmonic is <a href="https://x.com/satnam6502/status/1994094137386344665">hiring</a></h1><p>$120M series C.</p><p>Hardware is an interesting product area! Looks like their business model has advanced past the &#8220;mumbling to investors about curryhoward&#8221; stage. 2025, the year of mumbling to investors about curryhoward, has come to a roar of a close. I have also <a href="https://newsletter.for-all.dev/p/november-december-2024-progress-in">mumbled about curryhoward to my dearest yall</a>, which might mean I get bayes points for a math company starting to spin up a program synthesis product. I can&#8217;t tell how obvious that sort of claim was, or is, but I know one thing: <em>I love getting points</em>.</p><p>If you have Aristotle access, please test <a href="http://alokbeniwal@gmail.com">FVAPPS</a> and report back. Be sure to append the unit tests, that&#8217;s like the hardest part of the benchmark.</p><h1>If I had a nickel for every benchmark prefixed &#8220;Veri-&#8221; it&#8217;d only be four nickels but it&#8217;s still weird that it happened four times</h1><p>Some of these I had no good reason not to cover earlier. Abstracts styletransferred by Gemini.</p><h2><a href="https://arxiv.org/abs/2509.22908">Vericoding</a></h2><blockquote><p>We do hereby present and test the <strong>largest ledger of trials yet assembled</strong> for the craft known as <strong>Vericoding</strong>&#8212;the generation of a code whose certainty is sworn upon by the very stars&#8212;from the <strong>formal scrolls of specification.</strong> This, mind you, is in stark contrast to the common, wicked <strong>Vibe Coding,</strong> which spews forth a quick but bug-ridden script, born of a mere whisper of natural tongue.</p><p>Our grand ledger contains twelve thousand five hundred and four such scrolls of specification, with three thousand and twenty-nine written in the ancient runes of <strong>Dafny,</strong> two thousand three hundred and thirty-four in the sturdy tongue of <strong>Verus/Rust,</strong> and seven thousand one hundred and forty-one in the subtle logic of <strong>Lean.</strong> Of these, a full six thousand one hundred and seventy-four are entirely new, untarnished challenges.</p><p>We find that the success rate of this noble Vericoding, when performed by the <strong>Sorcerers of Language</strong> (our off-the-shelf LLMs), stands at a meager <strong>27% in Lean</strong>, rises to <strong>44% in Verus/Rust</strong>, and achieves a triumphant <strong>82% in Dafny.</strong> Alas, the addition of a common, flowery natural-tongue description does not notably sharpen their success. Furthermore, the light of these Sorcerers has illuminated the pure path of Dafny verification, raising its former success rate from a humble 68% to a glorious <strong>96%</strong> over the past twelve moons.</p></blockquote><h2><a href="https://openreview.net/forum?id=rWkGFmnSNl">Veribench</a></h2><blockquote><p><strong>The Formal Verification of Software</strong> doth stand as a promise most bright&#8212;a potential transformation wrought by the <strong>Generative Artifice of the Mind</strong> (AI). For a <strong>Provably Correct Code</strong> would utterly banish entire legions of hidden vulnerabilities, staunch the fatal breaches of critical systems, and, perhaps, forever change the practice of software engineering through trustworthy methods of implementation.</p><p>To spur this sacred domain, we unveil <strong>VeriBench,</strong> a trial meticulously crafted for judging the strength of the Sorcerers&#8217; Models in the <strong>end-to-end verification of the Code.</strong> This task demands the generation of complete Lean 4 incantations&#8212;the working functions, the unit tests, the <strong>Theorems of Correctness,</strong> and the <strong>Formal Proofs themselves</strong>&#8212;all drawn from humble Python reference spells or their accompanying common-tongue docstrings.</p><p>Our scrutiny of this one hundred and thirteen-task suite (comprising the tasks of HumanEval, simple drills, classical algorithms, and security snares) reveals a woeful truth: the current Frontier Sorcerers compile but a small fraction of the programs. <strong>Claude 3.7 Sonnet</strong> achieves compilation on a mere <strong>12.5%,</strong> while the mighty <strong>LLaMA-70B</strong> cannot compel a single program to compile in the Lean 4 HumanEval subset, even after fifty attempts guided by feedback! Yet, observe the noble <strong>Self-Optimizing Trace Agent</strong> architecture, whose compilation rates approach a magnificent <strong>60%!</strong> VeriBench thus lays the unyielding stone for developing systems capable of synthesizing provably correct, bug-free code, thus advancing the journey toward a more secure and dependable digital kingdom.</p></blockquote><h2><a href="https://arxiv.org/abs/2505.19271">VerifyThisBench</a></h2><blockquote><p>While the <strong>Grand Language Models</strong> (LLMs) have shown marvelous cunning in the quick generation of code, many existing trials are now easily conquered, and offer little <strong>guarantee of trustworthiness</strong> for the generated programs. To gain greater insight into the Sorcerers&#8217; reasoning on matters of <strong>Formal Correctness,</strong> we present <strong>VerifyThisBench,</strong> a new, agonizing trial which assesses the end-to-end verification of programs from mere natural-tongue descriptions.</p><p>The models must complete a trifecta of chivalric deeds: (i) <strong>Extract the Formal Specifications,</strong> (ii) <strong>Implement the Code</strong> in a language that craves verification, and (iii) <strong>Construct the Machine-Checkable Proofs.</strong></p><p>Our evaluation reveals that even the most vaunted of the modern models, such as <strong>o3-mini,</strong> achieve a pass rate of less than <strong>4%,</strong> with many of their utterances failing to even compile! To divine the true source of this difficulty, we further propose <strong>VerifyThisBenchXS,</strong> a milder variant where partial implementations or proofs are benevolently supplied. Across nine distinct models and seven tools of verification, we observe a steady gain when refinement is driven by the whispers of feedback, yet the overall pass rates remain pitifully low, underscoring the vast chasms that yet divide the Sorcerers from true formal reasoning. We release this trial and its unified environment to spur on the verification powers of all future models.</p></blockquote><h2><a href="https://arxiv.org/abs/2510.06296">VeriEquivBench</a></h2><blockquote><p><strong>Formal Verification</strong> stands as the ultimate frontier for ensuring the veracity of the code spawned by the <strong>Grand Language Models</strong> (LLMs). Methods that co-generate the code and the formal specifications in austere formal languages, such as Dafny, can, in theory, swear upon the truth of their alignment with the user&#8217;s intent. Alas, the entire progress is stifled by the <strong>difficulty of judging the quality of the specifications themselves.</strong></p><p>Current trials rely upon the perilous task of matching the generated work against a ground-truth specification&#8212;a manual process requiring deep expertise, which has limited existing datasets to a mere few hundred simple problems, and moreover suffers from a profound lack of reliability.</p><p>To remedy this, we introduce <strong>VeriEquivBench,</strong> a new trial featuring two thousand three hundred and eighty-nine complex algorithmic puzzles designed to expose the frailty of current models in both the generation of code and the deep formal reasoning. Our evaluative framework replaces the perilous ground-truth matching with a <strong>formally grounded metric: the Equivalence Score,</strong> and rigorously verifies the quality of the generated specifications and code. Our findings declare that the generation of formally verifiable code remains a <strong>profound challenge</strong> for the state-of-the-art Sorcerers. This underscores both the sheer difficulty of the task and the desperate need for trials like VeriEquivBench to hasten the march toward scalable and trustworthy coding agents.</p></blockquote><h1>From Galois&#8217; blog</h1><h2><a href="https://www.galois.com/articles/specifications-dont-exist">Specifications don&#8217;t exist</a></h2><p>Should&#8217;ve been in last newsletter but slipped through the cracks.</p><p>We need words for the different pessimisms about FMxAI. I often talk about the world-spec gap or the world-spec problem (that formal methods don&#8217;t rule out sidechannel attacks). This post is about a different pessimism, the elicitation problem or the elicitation and validation problem. Someone should absolutely be funding an org to focus on elicitation and validation, it&#8217;s a turbo important part of the theory of change. Is anyone working on this?</p><h2><a href="https://www.galois.com/articles/claude-can-sometimes-prove-it">Lean and claude code</a></h2><p>Mike also has a technical post about vibecoding in Lean.</p><p>Pair it with these <a href="https://github.com/cameronfreer/lean4-skills">off the shelf &#8220;skills&#8221;</a> (a claude code feature that&#8217;s &#8220;just prompts with extra steps&#8221;).</p><h2><a href="https://www.galois.com/articles/generative-ai-for-rigorous-digital-engineering">Rigorous Digital Engineering</a></h2><p>What if proof engineering but too cheap to meter?</p><h1><a href="https://www.logicalintelligence.com/product">Oops i missed Logical Intelligence</a></h1><p>Should&#8217;ve covered these folks a while ago. Yes, it appears their clientele is crypto/defi, but I have a generally positive attitude about life and I don&#8217;t want to set my &#8220;days since snark incident&#8221; counter back to zero, so we will ignore that and focus on the little we can ascertain about their tech and their claims.</p><p>There are two parts to this, there&#8217;s the part of why/how exactly they believe what they believe about their Lean product, and the part of how their Noa agent (which is not paywalled, you can just install it on github) fits into my strategic worldview.</p><h2>Primitive screwheads: text-to-text. <a href="https://youtu.be/zdkqagOUaPM?si=uJkFgd6LUwa55p2m">My boomstick</a>: structural synthesis</h2><p>Logical Intelligence is not bullish on autoregressive text-to-text as a program synthesis paradigm. Like <a href="https://centaur.stanford.edu/PyPantograph/intro.html">Leni Aniva</a>, they think tree search (starting with MCTS) will beat LLMs in the fullness of time. The interesting part, with a very paywalled model that I can&#8217;t test, is if they&#8217;re right why isn&#8217;t Harmonic (or Morph or a frontier company or anyone else) scooping them? It&#8217;s the same thing I say when I look at <a href="https://higherorderco.com/">HOC</a>: yes, text-to-text is an uncivilized approach to program synthesis, but we haven&#8217;t welded structural synthesis with the <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">bitter lesson</a> yet, and I don&#8217;t expect to see the gains until we do. If it could be any other way, then we&#8217;d be living in the GOFAI Extended Cinematic Universe instead of the Prompts Extended Cinematic Universe. I could write down some loose ideas of things you could try (to achieve the welding), but I will not because I&#8217;m unconvinced the <a href="https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html">d/acc</a> case is actually the majority of the mass. I&#8217;m too concerned that Logical Intelligence, HOC, to some extent Leni are <em>right</em> about the superpower unleashed by structure-aware program synthesis and I don&#8217;t think we&#8217;re ready (as a QA/safety community, nor as a society).</p><h2>Analyzing codebases for vulnerabilities</h2><p>From their product page:</p><blockquote><p>Ordering an external audit is both very expensive and very time-consuming. Our AI tool, Noa, delivers regular feedback on your code&#8212;minutes for smaller codebases and tens of minutes for larger ones. This lets you get near-real-time insight into the most critical potential security risks at a fraction of the cost. Noa integrates with GitHub: simply add the Noa bot to your repository, and after each pull request you can request a dashboard showing potential risks across the entire repository, along with their likelihood of exploitation and severity ratings.</p></blockquote><p>I have a post coming out about this, but I think the sort of thing they&#8217;re trying to do here is an important part of the strategic outlook. Audits, cryptanalysis, cybersecurity consulting are an important area to automate if we&#8217;re going to know, with a finite proof synthesis budget, which components are the most critical to harden with proofs. To be clear, I have not used the product, I don&#8217;t have any codebases it&#8217;s a good fit for. But it&#8217;s a class of product I&#8217;m excited about, even (ugh) if it is (ew) for defi/crypto.</p><h1><a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai#opportunityseeds">Announcements from the first round of Mathematics for Safe AI Opportunity Space at ARIA</a></h1><p>Spot ole q doc somewhere on this page! Other highlights are the <a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai?cardId=hardware-level-ai-safety-verification#opportunityseeds">hardware verification team</a>, the <a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai?cardId=gflownet-steered-probabilistic-program-synthesis#opportunityseeds">GFLowNet/SynthStats team</a>, and the <a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai?cardId=sfbench#opportunityseeds">SFBench team</a>.</p><h1><a href="https://arxiv.org/abs/2508.15878">Scalable synthesis of theorem proving challenges in formal-informal pairs</a></h1><p>Apparently there was some twitter discourse about this paper but one of the discoursers was using a hidden profile. It&#8217;d be great to be more like a <a href="https://thezvi.substack.com/">Zvi</a> style newsletter full of twitter screenshots, that would just require me to log onto to twitter more, which like, no.</p><blockquote><p><strong>The Grand Confluence of Lean and the Scholarly Arts of Computation: A Fount of Trials for the Sorcerer&#8217;s Mind&#8211; </strong>The noble art of <strong>Formal Theorem Proving (FTP)</strong> hath risen as a cornerstone for judging the deep reasoning capabilities of the <strong>Grand Language Models</strong> (LLMs), enabling the automated verification of mathematical oaths upon a massive scale. Yet, the progress of this quest has been hindered by a scarcity of suitable archives, due to the high toll of manual curation and the lack of truly challenging dilemmas paired with verified correspondences between <strong>Formal Scroll</strong> and <strong>Informal Chronicle. </strong>We propose to tap into the wellspring of <strong>Theoretical Computer Science (TCS)</strong> as a boundless source of rigorous proof problems. Within this scholarly domain, the definitions of algorithms permit the automatic synthesis of an arbitrary number of complex <strong>Theorem-Proof pairs. </strong>We demonstrate this potent approach upon two realms of TCS: the <strong>Busy Beaver problems,</strong> which demand the proof of bounds upon a Turing Machine&#8217;s cessation of movement, and the <strong>Mixed Boolean Arithmetic problems,</strong> which entwine the logic of the mind with the rigor of number. Our framework automatically weaves these challenges, providing parallel specifications: the <strong>Formal Code (Lean4)</strong> and the <strong>Informal Narrative (Markdown),</strong> thus creating a scalable conduit for generating verified trials of proof. Scrutiny of the frontier models reveals <strong>substantial chasms</strong> in their automated theorem-proving prowess: while the champion <strong>DeepSeekProver-V2-671B</strong> achieves a noble <strong>57.5%</strong> success rate on the Busy Beaver challenges, its strength wanes, managing only <strong>12%</strong> on the Mixed Boolean Arithmetic puzzles. These findings illuminate the great difficulty of crafting long-form proofs, even for those problems whose computational verification is a mere trifle, thus showcasing the invaluable role of TCS realms in advancing the research of automated reasoning.</p></blockquote><h1><a href="https://airesilience.net/vision-cyberphysical">AI Resilience: cyberphysical systems</a></h1><p>Friend of the newsletter Nora Ammann published AI Resilience a little bit ago. The section on cyberphysical systems is relevant to us: it relies on secure (formally verified) program synthesis becoming cheap and accessible. Resilience is a flavor of <a href="https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html">defensive acceleration</a> that specifically targets the durable and structural resolution of vulnerabilities, vulnerabilities which get amplified by AI but which, if we&#8217;re diligent and hardworking, get ameliorated by AI as well.</p><h1><a href="https://www.arxiv.org/abs/2506.04592">Let&#8217;s formalize this step by step</a></h1><p>One time a friend asked me &#8220;why not just put the proof synthesis in the reasoning trace and the thing you&#8217;re writing the proof about (say, a program) in the final output&#8220;. And I was like, &#8220;...huh&#8221;. And I got as far as adding a few credits to my runpod account before getting pulled into other things. Little did I know, at exactly that moment, this team was hard at work!</p><blockquote><p><strong>A Proposal for Safe Passage: The Formal Verification of the Grand Sorcerers&#8217; Thoughts&#8211; </strong>The method of the <strong>Chain-of-Thought (CoT)</strong> prompting hath become the established ritual for coaxing forth the reasoning powers from the <strong>Grand Language Models</strong> (LLMs). Yet, to contain the <strong>hallucinations</strong> in these Chains&#8212;phantoms notoriously difficult to discern&#8212;the current remedial arts, such as the Process Reward Models (PRMs) or the Self-Consistency measures, operate as <strong>opaque boxes,</strong> offering no verifiable evidence for their judgments, thus perhaps limiting their true efficacy. To redress this failing, we draw inspiration from the ancient wisdom that <strong>&#8220;the gold standard for supporting a mathematical claim is to provide a proof.&#8221;</strong> We propose a retrospective, step-aware framework of <strong>Formal Verification</strong> which we title <strong>Safe. </strong>Rather than assigning arbitrary scores or marks, we strive to articulate the mathematical claims within the <strong>formal mathematical language of Lean 4</strong> at the conclusion of <strong>each reasoning step,</strong> and further provide formal proofs to definitively identify these hallucinations. We test our framework <strong>Safe</strong> across various models and mathematical archives, demonstrating a <strong>significant enhancement</strong> in their performance, while simultaneously offering <strong>interpretable and verifiable evidence</strong> for their passage. Furthermore, we propose <strong>FormalStep</strong> as a new trial for the correctness of step-by-step theorem proving, containing <strong>30,809 formal statements. </strong>To the best of our knowledge, our work represents the <strong>first valiant endeavor</strong> to utilize the formal mathematical language of Lean 4 for verifying the natural-tongue content generated by the LLMs, thereby aligning with the very reason these formal languages were created: to provide a <strong>robust and unshakeable foundation</strong> for the hallucination-prone proofs scribed by human hands.</p></blockquote><h1><a href="https://ulyssean.com/">Ulyssean website mission status: totally sick</a></h1><p>There&#8217;s honestly no Ulyssean update in this issue, but I stumbled upon their website and loved the graphic design!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">There are no benefits for paid subscriptions. A Funder of Undisclosed Provenance is backing the newsletter for 6 months. </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[July-October 2025 Progress in Guaranteed Safe AI]]></title><description><![CDATA[Hot FMxAI Summer grinds along. Many new founders on the scene. Thinking of rebranding the newsletter away from &#8220;guaranteed&#8221;.]]></description><link>https://newsletter.for-all.dev/p/july-october-2025-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/july-october-2025-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Wed, 22 Oct 2025 02:28:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/85ef96ac-848c-4d60-bcab-bb66955e4728_1080x1080.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Yall, I really do apologize for radio silence. It has mostly to do with breaking my ankle in three places, but I&#8217;m walking again.</p><p>This edition of the newsletter looks a bit more like movement happenings and announcements, which isn&#8217;t to say that there weren&#8217;t more papers or technical results I could&#8217;ve included, just that my mind wasn&#8217;t on them over the summer. I feel like I should be working on strategic clarity right now! Watch this space, etc.</p><h1><a href="https://luma.com/k075jrgl">Verilib launch/demo</a> on the 23rd (SOON) </h1><p>The flagship product of the <a href="https://www.beneficialaifoundation.org/">Beneficial AI Foundation</a> is publicly launching! More on this in next edition of the newsletter. </p><h1><a href="https://theorem.dev/blog/anthropic-bug-test/">Theorem blog post</a></h1><p>I remain turbo impressed with Theorem&#8217;s tech. Formal methods is quality assurance, and as every QA technician learns on the first day of QA School, property based tests are 80% of the value of a proof for 20% of the effort.</p><blockquote><p>Why combine proofs and PBTs? Standard PBTs are compute efficient but prone to gaps: developers miss edge cases, and rare bugs slip into production. Proofs solve this with modular steps that decompose and incorporate reasoning, revealing what cases we&#8217;re missing&#8212;but they&#8217;re computationally impractical.</p></blockquote><p>They&#8217;re a little coy with the details. Too coy really, I thought of omitting this from the newsletter but I&#8217;ve seen some demos from the team that aren&#8217;t public yet that make me bullish on the company, so I want to remind everyone to watch this space.</p><h1><a href="https://xinhuajian.wordpress.com/wp-content/uploads/2025/04/ape_bench_i-2.pdf">APE TOGETHER STRONG</a></h1><p>No reason this couldn&#8217;t have been covered in an earlier episode! I think it slipped through the cracks or I had too many similar papers that month.</p><p>I&#8217;m pretty interested in the version control problem, in the long game. Every Lean synthesis laborer in the prompt mines every day wiring up their agents, MCP servers, and <code>lean-toolchain</code> files knows that the high code velocity of Lean and Mathlib relative to fewness of tokens in the pretraining data is a major pain point, especially compared to the synthesis successes of larger-resourced languages. What does it look like if language model technology eventually scales to something less hobbled by this?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>To be clear, this question of version control is different from the interest in version control over in <a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai">Safeguarded/ARIA</a> land <a href="https://gsai.substack.com/i/155386915/structure-aware-version-control-via-observational-bridge-types-hiring">that was covered in a previous newsletter</a>. Over there (davidadia? davidadistan?), specifications and world models are a kind of <em>structured data</em>, and they want version control that (among other things) doesn&#8217;t just flatten critical specs to plaintext.</p><p>Moreover, I don&#8217;t think this is necessarily the reason ByteDance got interested in version control for proof synthesis. They may be thinking diff-aware / git history aware approach to Mathlib leads to more synthetic data and a better understanding (in the model) of how to add features, refactor, etc.</p><blockquote><p>Recent progress in large language models (LLMs) has shown promise in formal theorem proving, yet existing benchmarks remain limited to isolated, static proof tasks, failing to capture the iterative, engineering-intensive workflows of real-world formal mathematics libraries. Motivated by analogous advances in software engineering, we introduce the paradigm of Automated Proof Engineering (APE), which aims to automate proof engineering tasks such as feature addition, proof refactoring, and bug fixing using LLMs. To facilitate research in this direction, we present APE-Bench I, the first realistic benchmark built from real-world commit histories of Mathlib4, featuring diverse file-level tasks described in natural language and verified via a hybrid approach combining the Lean compiler and LLM-as-a-Judge. We further develop Eleanstic, a scalable parallel verification infrastructure optimized for proof checking across multiple versions of Mathlib. Empirical results on state-of-the-art LLMs demonstrate strong performance on localized edits but substantial degradation on handling complex proof engineering. This work lays the foundation for developing agentic workflows in proof engineering, with future benchmarks targeting multi-file coordination, project-scale verification, and autonomous agents capable of planning, editing, and repairing formal libraries.</p></blockquote><h1><a href="https://arxiv.org/abs/2410.19940">Cobblestone</a></h1><p>This was originally posted a whole year ago, but had a revision over the summer. In any case, I had missed it until now.</p><p>I basically consider this a neurosymbolic architecture. The <code>hammer</code> (an SMT driven proof search tool) is the symbolic, and the regular LLM parts you know and love are the neuro part. Obviously you can do more than prompt the agent &#8220;try using the hammer tactic sometimes if you get stuck, little buddy :) I believe in you&#8221;. Their divide-and-conquer is one of those things.</p><p>Standard (by now) CoT tricks and agent stuff. It&#8217;s further ahead than a couple papers I&#8217;ve seen in the Lean agent space that came out since then, so good for them.</p><blockquote><p>Formal verification using proof assistants, such as Coq, is an effective way of improving software quality, but requires significant effort and expertise. Machine learning can automatically synthesize proofs, but such tools are able to prove only a fraction of desired software properties. We introduce Cobblestone, a divide-and-conquer approach for proof synthesis. Cobblestone uses a large language model (LLM) to generate potential proofs, uses those proofs to break the problem into simpler parts, automatically identifies which of those parts were successfully proven, and iterates on the remaining parts to build a correct proof that is guaranteed to be sound, despite the reliance on unsound LLMs. We evaluate Cobblestone on four benchmarks of open-source Coq projects, controlling for training data leakage. Fully automatically, Cobblestone outperforms state-of-the-art non-LLM tools, and proves many theorems that other LLM-based tools cannot, and on many benchmarks, outperforms them. Each Cobblestone run costs only $1.25 and takes 14.7 minutes, on average. Cobblestone can also be used with external input, from a user or another tool, providing a proof structure or relevant lemmas. Evaluated with such an oracle, Cobblestone proves up to 58% of theorems. Overall, our research shows that tools can make use of partial progress and external input to more effectively automate formal verification.</p></blockquote><h1>More math startups, but I&#8217;m not gonna single anyone out</h1><p>As always, I&#8217;m out in the applebees parking lot informing everyone that the VC influx into formal methods (especially Lean) doesn&#8217;t turn into AI security gains by default. We need you and I to make sure that formal verification agents are directed at problems in real world AI security and safety. <em>Some</em> or perhaps <em>most</em> math companies are doing PR in 2025 so they can create software QA products in 2026. I talked to one, though, who is adamant they are not. They just want to solve math problems and aren&#8217;t terribly fussed with capturing more than 1% of the upside. I think these guys are the outlier: most can be expected to pivot to program synthesis in 2026. I&#8217;m happy to discuss operationalizations of this to take bets. See also <a href="https://www.lesswrong.com/posts/kq8CZzcPKQtCzbGxg/quinn-s-shortform?commentId=hKfKAbXgGChxBuXDG">this brief memo I just jotted down</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/p/july-october-2025-progress-in-guaranteed?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">I feel like if you have friends who want to keep their ear to the ground about this math automation stuff you should send this to them, that way I can corrupt them into AI security/safety slowly over time. </p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/p/july-october-2025-progress-in-guaranteed?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.for-all.dev/p/july-october-2025-progress-in-guaranteed?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p></p><p><a href="https://www.smbc-comics.com/comic/signal-4">SMBC recently hit the nail on the nose</a>. It feels in the direction of if I had ranted at Zach Weinersmith himself, but he went way harder and took it way further.</p><h1>I know a bunch of founders who are aiming to increase headcount soon</h1><p>And while they have not literally posted their job ads yet, reach out and I can let you know if I think they&#8217;d like to talk to you.</p><p>I&#8217;d love to do some napkin math about the VC influx in FMxAI over the last year, and include projections for how it might grow in the next couple years. I&#8217;d also love for a reader to do it and I can just link to it here!</p><h1><a href="https://luma.com/8bg27vqg?tk=kfEggv">Clark Barrett&#8217;s appearance on Orpheus&#8217; seminar series</a> is finally scheduled</h1><p>Separately, <a href="https://manifund.org/projects/guaranteed-safe-ai-seminars-2026-5kxriozt23?tab=comments">Orpheus has a Manifund post up to support the seminars</a> and his other GSAI movement building efforts. Early endorsements by both of the <a href="https://gsai.substack.com/i/158079474/fvapps">FVAPPS</a> coauthors.</p><h1>TA2 is dead over at ARIA</h1><p>I think we&#8217;ve finally had a newsletter published while ARIA does not have an active Safeguarded AI funding call. Achievement unlocked! The original TA2 plan to fund one team to the tune of 8 figures was canceled in the 9th inning.</p><blockquote><p>When we designed this programme, the world looked different. Today, the pace of progress in frontier AI models has fundamentally altered the path to our goals. We now expect that the intended technical objectives of TA2 will be attainable as a side effect of this progress, without requiring a dedicated R&amp;D organisation. Instead of investing in creating specialised AI systems that can use our tools, it will be more catalytic to broaden the scope and power of the TA1 toolkit itself, making it a foundational component for the next generation of AI.</p></blockquote><p>TA2, <a href="https://gsai.substack.com/i/161749219/ta-call">as you&#8217;ll recall</a>, was to be a large investment in founding a new org that would be kinda frontier-lab like in some ways but it would specialize in being superpowered at the new specification - world model - proof cert language/stack that TA1 is shipping. I think it&#8217;s roughly true that advanced capability in this new language/stack can be accomplished as a side effect of what claude, gpt, grok, and gemini are going to do anyway. But I&#8217;m also surprised that the extent to which that is forecasted to be true now, by davidad and his team, wasn&#8217;t priced in back when the Plan&#8482; was originally drawn up. Davidad just seems a little more bitter lesson / scaling pilled than being surprised enough by capabilities alone to pivot! I guess I have a mild impulse to speculate that the team got spooked by the difficulty of aligning incentives for this kind of org to get a 7 or 8 figure injection at its founding, but again I&#8217;d ultimately expect that to have been priced in when the Plan&#8482; was originally published.</p><p>I&#8217;ve heard of at least one turbo sick venture that was germinated by the prospect of pursuing this grant, and I don&#8217;t think they&#8217;re giving up just yet. Watch this space.</p><h1><a href="https://manifund.org/lisathiergart">Lisa from SL5 Taskforce</a> still up at Manifund</h1><p>Lisa from the <a href="https://securityandtechnology.org/sl5/">SL5 Taskforce</a> has a few more months to disperse up to $100k. SL5 Taskforce is interested in some applications of FMxAI, especially to cloud infrastructure hardening. The center of the venn diagram (between Progress in GSAI readers and the interests of the SL5 Taskforce) is not massive, but I think it&#8217;s big enough for us to discuss here!</p><p><a href="https://docs.google.com/forms/d/e/1FAIpQLScpNZfNuF1ByTXL1zJzBEuuGrTYpaNscBbQZs3Gt6DVt3bITw/viewform">Apply here</a>, I think you also have to file a Manifund project as well.</p><h1>Cool NASA <a href="https://ntrs.nasa.gov/citations/20250006044">slidedeck from the summer about formal methods at Langley</a></h1><p>I think the &#8220;future&#8221; section on cyberphysical systems and comparing AI to cyberphysical systems is good. The AI4FV section is all the stuff we talk about, and FV4AI sections are all the stuff I&#8217;m constantly saying we&#8217;re not talking about. The highlights are the earlier parts, the historical context, IMO.</p><h1>Rebranding the newsletter</h1><p>Yall, I&#8217;m interested in your thoughts here. My priorities are shaped a lot by &#8220;ensure AI security knows what to ask the formal methods community for at crunch time&#8221; and the converse &#8220;ensure that formal methodsititians view the AI security community as their most important customer&#8221;. I think GSAI is not completely wrong, for the title of the newsletter! But I think it could be even less wrong. The main problems I see with &#8220;guarantee&#8221; is that it doesn&#8217;t evoke swiss cheese, and I think formal methods are a source of swiss cheese!</p><p>I also have to disambiguate AI security as I mean it (boxing) from what e.g. <a href="https://palisaderesearch.org/">Palisade</a> means (offensive infosec capabilities). Watch this space etc.</p><p>Also if you want to help me with the newsletter we could be a team! I don&#8217;t think I could get back up to once per month without some friends to help.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">There are no benefits to being a paid subscriber. I&#8217;m also frequently late as heck to post, as evidenced by this issue being a 4-month roundup. The help is encouraging and meaningfully makes me keep doing this. Anyway, subscription without paying me just puts it in your inbox right when they come out.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Caveat: this problem was way worse 1-2 years ago. I think as the Lean3-&gt;Lean4 migration cooled down, velocity as apparent in pretraining data got lower, <em>plus</em> language models getting much more powerful. Overall I&#8217;m curious if this velocity/version control problem will remain persistent, not turbo confident it will.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I think this shortform is pretty important, I do regret being unnuanced about curryhoward in the past, separately my worldview has gotten more nuanced and changed in other ways. Or I think back when I said &#8220;something something curryhoward&#8221; more, being directionally correct was enough, but now that more people are saying it, it&#8217;s time to be not just directionally correct but precise. You get the idea.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[May-June 2025 Progress in Guaranteed Safe AI]]></title><description><![CDATA[Benchmarks! Get your benchmarks! A startup and a nonprofit lab each launch]]></description><link>https://newsletter.for-all.dev/p/may-june-2025-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/may-june-2025-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Thu, 26 Jun 2025 21:20:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Yb4J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There will be a AIxFM conference in the Bay Area Q4, according to a little birdie.</p><h1><a href="https://github.com/morph-labs/lean-abc-true-almost-always">Morph ships big autoformalization result in 3599 lines of Lean</a></h1><p>They have human decomposition in the latex/lean blueprint, into 67 lemmas with human spotchecking. Still, I&#8217;m impressed with their system (called Trinity).</p><p>I&#8217;d like to know how expensive (in tokens, or some other compute metric) it was to do this!</p><h2>On <a href="https://www.morph.so/blog/verified-superintelligence">Verified Superintelligence</a></h2><p>I of course have opinions on their blog post Verified Superintelligence.</p><blockquote><p>Today's most advanced AI systems&#8212;reasoning LLMs trained with supervised RL&#8212;have hit a fundamental wall. They can only improve on problems where we can verify the (known) answers. Every math problem needs a known solution. Every coding challenge requires test cases. Every reasoning task demands ground truth.</p></blockquote><p>I think I know where they&#8217;re going with this: &#8220;verification&#8221; in the sense of supervised RL as a cognate with &#8220;verification&#8221; in the sense of typechecking.</p><blockquote><p>Perhaps most importantly, this approach offers a path to what we call "trustless alignment." By ensuring all outputs can be independently validated, we create AI systems whose safety doesn't depend on faith in their training process or architecture. The proof is in the output.</p></blockquote><p>This describes effectively a <a href="https://www.lesswrong.com/w/ai-boxing-containment">boxing</a> strategy, not alignment. Boxing is when you police the boundary between the AI and the world. Alignment is when you don&#8217;t need to. Is this an important distinction, or a narcissism of small differences you&#8217;d only notice if you have Lesswrong Disease (the disease they diagnose you with when you read or write on Lesswrong)? Being afflicted myself, I&#8217;m not sure how I would tell.</p><h1>Two benchmarks</h1><h2><a href="https://arxiv.org/abs/2505.23135">VERINA</a></h2><blockquote><p>Large language models (LLMs) are increasingly integrated in software development, but ensuring correctness in LLM-generated code remains challenging and often requires costly manual review. Verifiable code generation -- jointly generating code, specifications, and proofs of code-specification alignment -- offers a promising path to address this limitation and further unleash LLMs' benefits in coding. Yet, there exists a significant gap in evaluation: current benchmarks often lack support for end-to-end verifiable code generation. In this paper, we introduce Verina (Verifiable Code Generation Arena), a high-quality benchmark enabling a comprehensive and modular evaluation of code, specification, and proof generation as well as their compositions. Verina consists of 189 manually curated coding tasks in Lean, with detailed problem descriptions, reference implementations, formal specifications, and extensive test suites. Our extensive evaluation of state-of-the-art LLMs reveals significant challenges in verifiable code generation, especially in proof generation, underscoring the need for improving LLM-based theorem provers in verification domains. The best model, OpenAI o4-mini, generates only 61.4% correct code, 51.0% sound and complete specifications, and 3.6% successful proofs, with one trial per task. We hope Verina will catalyze progress in verifiable code generation by providing a rigorous and comprehensive benchmark. We release our dataset on <a href="https://huggingface.co/datasets/sunblaze-ucb/verina">this https URL</a> and our evaluation code on <a href="https://github.com/sunblaze-ucb/verina">this https URL</a>.</p></blockquote><h2><a href="https://arxiv.org/abs/2505.13938">CLEVER</a></h2><blockquote><p>We introduce CLEVER, a high-quality, curated benchmark of 161 problems for end-to-end verified code generation in Lean. Each problem consists of (1) the task of generating a specification that matches a held-out ground-truth specification, and (2) the task of generating a Lean implementation that provably satisfies this specification. Unlike prior benchmarks, CLEVER avoids test-case supervision, LLM-generated annotations, and specifications that leak implementation logic or allow vacuous solutions. All outputs are verified post-hoc using Lean's type checker to ensure machine-checkable correctness. We use CLEVER to evaluate several few-shot and agentic approaches based on state-of-the-art language models. These methods all struggle to achieve full verification, establishing it as a challenging frontier benchmark for program synthesis and formal reasoning. Our benchmark can be found on GitHub (<a href="https://github.com/trishullab/clever">this https URL</a>) as well as HuggingFace (<a href="https://huggingface.co/datasets/amitayusht/clever">this https URL</a>). All our evaluation code is also available online (<a href="https://github.com/trishullab/clever-prover">this https URL</a>).</p></blockquote><p>Welcome to the summer of benchmarks. VERINA and CLEVER each critically cite <a href="https://gsai.substack.com/i/158079474/fvapps">a paper we saw in a previous newsletter</a> to patch some of its issues, especially around completeness of specification synthesis.</p><h1><a href="https://www.ycombinator.com/launches/NZA-theorem-ai-coding-that-is-trustworthy-by-default">The Theorem launch</a></h1><p>Friends of the newsletter Jason and Rajashree launched Theorem (YC W25)</p><blockquote><p>MISSION: Theorem is an AI and programming languages research lab. Our question: as the models get better at generating code, what is the bottleneck to safely deploying vastly more computation in the world?</p><ul><li><p>Code has bugs. AI-enabled attackers will increase the volume and velocity of attacks on software infrastructure. Human code review will not scale.</p></li><li><p> AIs are untrusted generators. In order to deploy AIs in critical systems without fully understanding their generalization behavior, we&#8217;ll need robust, scalable methods for overseeing their solutions.</p></li></ul></blockquote><p>I&#8217;ve seen some of the tech and it&#8217;s fantastic; when I think about projects I could do, I try only to do things that Theorem wouldn&#8217;t be able to do better.</p><h1><a href="https://web.cs.ucdavis.edu/~cdstanford/doc/2024/ASEW24b.pdf">Proof Carrying Code Completions</a></h1><p>Would&#8217;ve been great to have covered this paper when it came out, but it&#8217;s relevant enough to discuss it even though I&#8217;m late.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Yb4J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Yb4J!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Yb4J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg" width="566" height="441" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:441,&quot;width&quot;:566,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Yb4J!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Yb4J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60dfd380-e6a1-465a-b32b-7cb8b480ba36_566x441.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p>Code completions produced by today&#8217;s large language models (LLMs) offer no formal guarantees. We propose proof-carrying code completions (&#119875;&#119862;^3). In this paradigm, a high-resourced entity (the LLM provided by the server) must provide a code completion together with a proof of a chosen safety property which can be independently checked by a low-resourced entity (the user). In order to provide safety proofs without requiring the user to write specifications in formal logic, we statically generate preconditions for all dangerous function calls (i.e., functions that may violate the safety property) which must be proved by the LLM.</p><p>To demonstrate the main ideas, we provide a prototype implementation in the program verification language Dafny, and a case study focusing on file system vulnerabilities. Unlike Python code generated by GPT-4, Dafny code generated by &#119875;&#119862;^3 provably avoids a common weakness related to path traversal (CWE-35), using a single generation attempt (&#119896; = 1) and a modest number of tokens (3, 350). Our tool is available as an open source repository at <a href="https://github.com/DavisPL/PCCC">https://github.com/DavisPL/PCCC</a>.</p></blockquote><p>The setup in this paper will be familiar to people who&#8217;ve been thinking about guaranteed safe AI for a bit. We have trusted and untrusted components, where a core trusted component is a proof checker of some kind.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CT6P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CT6P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 424w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 848w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 1272w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CT6P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png" width="1090" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1090,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CT6P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 424w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 848w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 1272w, https://substackcdn.com/image/fetch/$s_!CT6P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed29f4cf-5729-4992-b45f-7678106e63e7_1090x600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://en.wikipedia.org/wiki/Proof-carrying_code">Proof carrying code</a> is conceptually similar to a sigma type or dependent pair. It just means you tag or decorate whatever code you deliver with a proof of its correctness.</p><h1><a href="https://lawzero.org/en">Law Zero</a> launch</h1><p>Yoshua Bengio&#8217;s new org, pursuing the <a href="https://yoshuabengio.org/2024/02/26/towards-a-cautious-scientist-ai-with-convergent-safety-bounds/">Scientist AI</a> agenda. Scientist AI is to an extent in the same conceptual family as Safeguarded AI. Both are heavily involved in probability theory, where Scientist AI has even more information theory. Both are trying to avoid informally specified agents running amok, both by locking AI behind some probabilistic assurances. Safeguarded AI, however, emphasizes a robust notion of <em>certificate</em>, while Scientist AI doesn&#8217;t emphasize this. People have criticized Scientist AI for falling into the pitfalls of &#8220;oracles&#8221; that Bostrom et al discussed over a decade ago. Specifically, agency can&#8217;t be blocked without banning <em>loops</em> and subprocess.run, since tool use can be easily recovered in this day and age.</p><p>Law Zero is <a href="https://job-boards.greenhouse.io/lawzero">hiring</a> for sr roles in Montreal.</p><h1><a href="https://aisecurity.forum/">AI Security Forum in Tel Aviv, DC, and Vegas this summer</a></h1><p>Following the success of Defcon satellite events and the Paris AI Action Summit satellite event, the security forum hits Tel Aviv and DC for the first time before returning to Vegas.</p><h1><a href="https://foresight.org/request-for-proposals/security-technologies-for-securing-ai-systems/">Foresight RFP</a> alludes to some GSAI topics</h1><blockquote><ul><li><p>AI tools that assist with increasing the tractability of formal verification for real-world systems, starting at security-critical infrastructure. </p></li><li><p>Designing or proving properties of secure system architectures.</p></li></ul></blockquote><p>They accept applications rolling year round, but evaluate and disperse on a quarterly basis. The next &#8220;deadline&#8221; is June 30th. Sorry I didn&#8217;t tell you sooner, I&#8217;m bad at my job.</p><h1><a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai/funding">Safeguarded AI Technical Area 2</a></h1><p>Time to fire up a machine learning lab to automate all the proof engineering in the new proof stack that TA1 cooks up. Deadline October 1st.</p>]]></content:encoded></item><item><title><![CDATA[March-April 2025 Progress in Guaranteed Safe AI]]></title><description><![CDATA[Who even pays for patching open source software? And ARIA news]]></description><link>https://newsletter.for-all.dev/p/march-april-2025-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/march-april-2025-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Sun, 20 Apr 2025 18:53:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Say hi at ICSE in Ottawa, I&#8217;ll be at the reception Thursday, <a href="https://lu.ma/umi3g2wc?tk=6vGnXV">this colocated event on Friday</a>, and the <a href="https://llm4code.github.io/">LLM4Code workshop on Saturday</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DfMS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DfMS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DfMS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DfMS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DfMS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd314091c-6306-4482-b923-c640ea5323f4_1600x900.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As usual there are no benefits to the paid subscription. </p><p>Sorry for consolidating two months into one post again after I said I wouldn&#8217;t. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.for-all.dev/subscribe?"><span>Subscribe now</span></a></p><h1><a href="https://forum.effectivealtruism.org/posts/sh4zDoyQiwnAGCRd6/patching-all-security-relevant-open-source-software">Fermstimate of the cost of patching all security relevant open source software</a></h1><p>Niplav writes</p><blockquote><p>So, <em>a proposal</em>: Whenever someone claims that LLMs will d/acc us out of AI takeover by fixing our infrastructure, they will also have to specify who will pay the costs of setting up this project and running it.</p></blockquote><p>I&#8217;m almost centrally the guy claiming LLMs will d/acc us out of AI takeover by fixing infrastructure, technically I&#8217;m usually hedging more than that but it&#8217;s accurate in spirit.</p><blockquote><p>If transformative AI is developed soon, most open source projects (especially old ones relevant to internet infrastructure) are going to be maintained by humans with human response times. That will significantly increase the time for relevant security patches to be reviewed and merged into existing codebases, especially if at the time attackers will submit AI-generated or co-developed subtle exploits using AI systems six to nine months behind the leading capabilities, keeping maintainers especially vigilant.</p></blockquote><p>I usually say we prove the patches correct! But Niplav is correct: it&#8217;s a hard social problem, many critical systems maintainers are particularly slop-phobic and won&#8217;t want synthetic code checked in. That&#8217;s why I try to emphasize that <strong>the two trust points are the spec and the checker</strong>, and the rest is relinquished to a shoggoth. That&#8217;s the vision anyway&#8211; we solve this social problem by involving the slop-phobic maintainers in writing the spec and conveying to them how trustworthy the deductive process is.</p><p><a href="https://squigglehub.org/models/niplav/automated_internet_hardening">Niplav&#8217;s squiggle model</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7r8r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7r8r!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 424w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 848w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 1272w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7r8r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png" width="1456" height="572" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:572,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7r8r!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 424w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 848w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 1272w, https://substackcdn.com/image/fetch/$s_!7r8r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66f77b08-e2cd-44b2-a3a7-671d1199eb27_1600x629.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Median $~1b worth of tokens, plus all the &#8220;setting up the project, paying human supervisors and reviewers, costs for testing infrastructure &amp; compute, finding complicated vulnerabilities that arise from the interaction of different programs&#8230;&#8221; etc costs. I think a lot&#8217;s in our action space to reduce those latter costs, but the token cost imposes a firm lower bound.</p><p>But this is an EA Forum post, meaning the project is being evaluated as an EA cause area: is it cost effective? To be cost effective, the savings from alleviating some disvalue have to be worth the money you&#8217;ll spend. As a programming best practices chauvinist, one of my pastimes is picking on CrowdStrike, so let&#8217;s not pass up the opportunity. The <a href="https://en.wikipedia.org/wiki/2024_CrowdStrike-related_IT_outages#Impact">2024 outage is estimated to have cost</a> about $5b across the top 500 companies excluding microsoft. A public goods project may not have been able to avert CrowdStrike, but it&#8217;s instructive for getting a flavor of the damage, and this number suggests it could be easily worth spending around Niplav&#8217;s estimate. On cost effectiveness though, even I (who works on this &#8220;LLMs driving Hot FV Summer&#8221; thing full time) am skeptical, only because open source software is pretty hardened already. Curl/libcurl saw 23 CVEs in 2023 and 18 in 2024, which it&#8217;d be nice to prevent but really isn&#8217;t a catastrophic amount. Other projects are similar. I think a lot about the Tony Hoare quote &#8220;It has turned out that the world just does not suffer significantly from the kind of problem that our research was originally intended to solve.&#8221; Not every bug is even an exploit.</p><h1>Safeguarded AI</h1><h2><a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai/technical-areas">Creator profiles for TA1.1 and TA3</a></h2><p>The TA1.1 ones have been up for a while and I think I missed them, but TA3 was posted fairly recently.</p><p>TA3 is the applications domain. Yaknow, the real world, that thing from all those functional programming jokes.</p><blockquote><p>This work aims to address the challenge of AI deployment in safety-critical sectors by combining scientific world models and mathematical proofs to develop robust safety assurances.</p><p>Researchers from diverse companies and organisations such as AstraZeneca, Mind Foundry, the University of Oxford, the University of Birmingham, and the University of York, will aim to demonstrate how safeguarded AI systems can help them to make significant improvements in vital sectors.</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UJ8u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UJ8u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 424w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 848w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 1272w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UJ8u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png" width="1456" height="1540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e17018dd-2942-4205-a617-712a60556fbf_1513x1600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1540,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UJ8u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 424w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 848w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 1272w, https://substackcdn.com/image/fetch/$s_!UJ8u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe17018dd-2942-4205-a617-712a60556fbf_1513x1600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/safeguarded-ai/funding">TA2 call</a></h2><p>This isn&#8217;t really open to individuals, but as always reading the funding call gets you some insight into ARIA&#8217;s worldview and strategy</p><blockquote><p>ARIA is launching a multi-phased solicitation for Technical Area 2 (TA2) to support the development of a general-purpose Safeguarded AI workflow. The programme aims to demonstrate that frontier AI techniques can be harnessed to create AI systems with verifiable safety guarantees. In TA2, we will award &#163;18m to a non-profit entity to develop critical machine learning capabilities, requiring strong organizational governance and security standards. Phase 1, backed by &#163;1M, will fund up to 5 teams to spend 3.5 months to develop full Phase 2 proposals. Phase 2 &#8212; which will open on 25 June 2025 &#8212;will fund a single group, for &#163;18M, to deliver the research agenda. TA2 will explore leveraging securely-boxed AI to train autonomous control systems that can be verified against mathematical models, improving performance and robustness. The workflow will involve forking and fine-tuning mainstream pre-trained frontier AI models to create verifiably safeguarded AI solutions.</p></blockquote><p>I&#8217;ve heard this described as something like a &#8220;new frontier lab&#8221;, which is &lt;sarcasm&gt;always music to my ears&lt;/sarcasm&gt;, but in truth I&#8217;m only worried about shops that push the envelope on scaling (chinchilla or otherwise). Boromir wanted to use the One Ring to defend Gondor, and we largely trust Tolkien/whomever when they say that defensive acceleration is folly and the <s>Ring&#8217;s corrupting influence</s> incentives are stronger than the will of man. Nevertheless, I feel pretty good about this because it&#8217;s narrow and targeted. Eager to find out if I&#8217;ll eat my words.</p><h1>Conferences, events</h1><ul><li><p><a href="https://sos-vo.org/group/hcss_conference/cfp">High Confidence Software and Systems Conference</a> May 12, Maryland US</p></li><li><p><a href="https://conferences.i-cav.org/2025/">CAV</a> July 21, Croatia</p></li><li><p><a href="https://lu.ma/umi3g2wc?tk=6vGnXV">Formal methods for Rust</a> May 2nd, Ottawa CA</p></li></ul><h1><a href="https://arxiv.org/abs/2504.10374">New control paper</a></h1><p>GSAI and Control are really complimentary forces so I think I&#8217;ll mention news in Control, at least sometimes</p><blockquote><p>Control evaluations measure whether monitoring and security protocols for AI systems prevent intentionally subversive AI models from causing harm. Our work presents the first control evaluation performed in an agent environment. We construct BashBench, a dataset of 257 challenging multi-step system administration tasks, and evaluate whether various safety measures can prevent an adversarially constructed AI agent from covertly downloading and executing malicious code in this environment. This multi-step setting introduces new attack and defense dynamics, which we investigate in order to design novel control protocols that prevent safety failures without hindering the ability of non-malicious agents to perform useful work. We introduce a class of control protocols called resample protocols that dynamically take additional samples of certain actions. We find these protocols significantly improve on existing techniques by selectively blocking the AI agent from executing suspicious code and incriminating the agent by generating additional examples of dangerous behavior. We measure the tradeoff between attack prevention and usefulness; our best protocol combines resampling with analysis of previous steps, reducing the success rate of attacks from 58% to 7% at a 5% cost to the performance of a non-malicious agent.</p></blockquote>]]></content:encoded></item><item><title><![CDATA[January-February 2025 Progress in Guaranteed Safe AI]]></title><description><![CDATA[Proving the Coding Interview, R1, a job, and yet another ARIA solicitation]]></description><link>https://newsletter.for-all.dev/p/january-february-2025-progress-in</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/january-february-2025-progress-in</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Fri, 28 Feb 2025 03:04:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Kogt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Ok this one got too big, I&#8217;m done grouping two months together after this.</p><p>BAIF wants to do user interviews to prospect formal verification acceleration projects, reach out if you&#8217;re shipping proofs but have pain points!</p><p>This edition has a lot of my takes, so I should warn you that GSAI is a pretty diverse field and I would expect all my colleagues in it to have substantial disagreement with at least one take in each issue of the newsletter. Prompt the language model to insert the standard boilerplate disclaimer that views are my own, not those of my colleagues, my employer nor my mom.</p><p>If you're just joining us,<a href="https://arxiv.org/abs/2405.06624"> background on GSAI here</a>.</p><h1>Formal Verification is Overrated (lightning talk)</h1><p><a href="https://www.youtube.com/watch?v=bs5snugP1VA">Zac makes three sound points</a> which are not cruxes for me. Part of what <a href="https://www.lesswrong.com/posts/DZuBHHKao6jsDDreH/in-response-to-critiques-of-guaranteed-safe-ai">Nora&#8217;s post is doing</a> is responding to this with her take, which differs from mine.</p><h2>Point one: weights are intractable</h2><p>Formal verification of the learned component itself is gonna be a nonstarter on nontoy problems for computational cost reasons. <a href="https://gsai.substack.com/i/147012882/guarantees-based-mechanistic-interpretability-team-preprint">We saw last year</a> that a &#8220;proof&#8221; of the correctness of a function a transformer implements is subcubic in d_vocab, at least a little worse than quadratic. I agree with Zac, which is why I avoid <em>whitebox</em> strategies in my day to day life. GSAI (to me) is a set of <em>blackbox</em> tactics, closer to <a href="https://openreview.net/forum?id=KviM5k8pcP">control</a>, that doesn&#8217;t depend on interpretability going particularly well. The upcoming Formalise track of ICSE separates &#8220;AI4FV&#8221; and &#8220;FV4AI&#8221; as two separate topic areas, which is obviously the right thing to do and I find half of the time when I introduce GSAI to someone this is the first point of clarification. Speaking for myself, GSAI is AI4FV without being FV4AI.</p><h2>Point two: reality is complicated and unknown</h2><p>This <a href="https://www.lesswrong.com/posts/Egw5M6cnwgpRBXAQ4/speedrunning-4-mistakes-you-make-when-your-alignment#Mistake_1__The_gap_between_the_world_and_the_spec_hurts_you_more_than_the_gap_between_the_spec_and_the_implementation">capitulates a point I&#8217;ve been making for years</a>, which is that the world-spec gap hurts you more than the spec-implementation or spec-component gap. I think if we get arbitrary wins on GSAI agendas, but an arbitrarily unboxing superintelligence too quickly, this principle is what kills us. This principle (<a href="https://www.lesswrong.com/posts/kq8CZzcPKQtCzbGxg/quinn-s-shortform?commentId=5YhyYC7oNFWSmLPWe">which needs a good name, btw</a>) still allows you to stack your <a href="https://en.wikipedia.org/wiki/Defense_in_depth_(computing)">swiss cheese</a> tower to the sky, and never says that your swiss cheese tower buys you <em>nothing</em>. A potential crux for me might be that if we leave our immediate successors with a tower of swiss cheese, they&#8217;ll fail to use that to escape the acute risk period for predictable reasons.</p><p>Agree that the DNA synthesis part of <a href="https://arxiv.org/abs/2309.01933">Tegmark and Omuhundro</a> was a bit unhinged. Agree that insofar as Zac is responding to that paper, much &#8220;calming down&#8221; is needed cuz that paper (unlike, I think, the <a href="https://arxiv.org/abs/2405.06624">Toward GSAI</a> paper) is really overstated.</p><p>Formal verification, as we know it, exists to <strong>isolate error to the spec-world gap</strong>, which realistically isolates attack surfaces to a fraction of that. This doesn&#8217;t happen for free by default, it requires defensive acceleration. My job is <em>nines</em>, I don&#8217;t believe in 100%. However, a significant part of Safeguarded AI is aimed at making the spec-world gap <em>smaller</em>, so it wouldn&#8217;t be accurate to say that GSAI <em>stops</em> at the spec-world gap. I think Safeguarded AI&#8217;s approach to this, which attacks the details of probabilistic world models and how they&#8217;re elicited and falsified, is more reasonable than Omohundro&#8217;s &#8220;<a href="https://www.lesswrong.com/posts/P8XcbnYi7ooB2KR2j/provably-safe-ai-worldview-and-projects#Some_Useful_Projects">formal specifications of physical systems</a>&#8221; approach which is a little too bottom-up for my tastes.</p><p>I did talk to one security researcher who, after regaling me with fun tales of side channel attacks, said Omohundro&#8217;s vision of eliminating sidechannel attacks by formal verification because the model of the component goes down to physical laws like Maxwell&#8217;s equations is not as unreasonably intractable as I thought and is instead merely ambitious. So maybe I should learn basic physics.</p><h2>Point three: tool AI is unstable and uncompetitive</h2><p>Zac:</p><blockquote><p>The final category of proposals I see under this formal verification approach is that we should use AI to develop tools and software which is formally verified. And I am broadly a fan of this, but it&#8217;s not actually a substitute for working on AI safety.</p></blockquote><p>Also Zac:</p><blockquote><p>The addition of a trivial for loop is the only difference between a tool and an agent (in a general sense).</p></blockquote><p>I think Zac is referring to <a href="https://blog.biocomm.ai/2024/12/01/why-we-should-build-tool-ai-not-agi-max-tegmark-at-websummit-2024-future-of-life-institute/">the ask to restrain AI to tool level and </a><em><a href="https://blog.biocomm.ai/2024/12/01/why-we-should-build-tool-ai-not-agi-max-tegmark-at-websummit-2024-future-of-life-institute/">simply </a></em><a href="https://blog.biocomm.ai/2024/12/01/why-we-should-build-tool-ai-not-agi-max-tegmark-at-websummit-2024-future-of-life-institute/">not build agents</a>, coming mostly from Tegmark. (See <a href="https://futureoflife.org/wp-content/uploads/2025/02/AI-Action-Summit-Tool-AI-Explainer-V5.pdf">FLI&#8217;s tiers</a>. it&#8217;s possible that scaffolding more than a loop is all you need to get from tier 2 to tier 3). I just don&#8217;t think this ask is centrally related to formal verification, and is not a crux for most of the GSAI space (indeed, much of GSAI wants to leverage primitive agents such as loops). But I do agree with Zac, I think the no-agency ship has sailed and there&#8217;s no satisfying way to restrict AI to the tool level, most for competitiveness reasons.</p><h2>Overrated by whom?</h2><p>But also, I&#8217;d like to ask Zac how it&#8217;s &#8220;overrated&#8221; when the reception from funders (besides ARIA and FLI&#8217;s vitalikbucks) is not even lukewarm. OpenPhil is super unplugged from GSAI and not interested, SFF <a href="https://survivalandflourishing.fund/sff-2024-recommendations">isn&#8217;t participating</a> at all in spite of having some of <a href="https://arxiv.org/abs/2405.06624">the big position paper</a> authors involved in the S-process. Frontier labs do a bunch of proof scaling projects, but they&#8217;re doing it just for fun / to show off / to innovate on the product, not because of any vision in particular (as far as I can tell / as far as I&#8217;ve heard). I think ARIA and the FLI Extended Cinematic Universe (i.e. BAIF) is a great place to start, we could spend more but we shouldn&#8217;t complain. Does Zac mean the current level of funding is already too high, or is he just worried about that number increasing? He does think it&#8217;s an important part of the portfolio, so I think he&#8217;s just responding to some of the over-promising (which I tend to agree with him about).</p><h1>FVAPPS</h1><p>4715 function signatures each with 2-5 sorry&#8217;s out theorems: a new benchmark for leetcode-style programming in Lean. Notice that I, Quinn, the newsletter author, is one of the authors.</p><blockquote><p><a href="https://arxiv.org/abs/2502.05714">We introduce</a> the Formally Verified Automated Programming Progress Standards, or FVAPPS, a benchmark of 4715 samples for writing programs and proving their correctness, the largest formal verification benchmark, including 1083 curated and quality controlled samples. Previously, APPS provided a benchmark and dataset for programming puzzles to be completed in Python and checked against unit tests, of the kind seen in technical assessments in the software engineering industry. Building upon recent approaches for benchmarks in interactive theorem proving, we generalize the unit tests to Lean 4 theorems given without proof (i.e., using Lean&#8217;s &#8220;sorry&#8221; keyword).</p></blockquote><p>Workshop reviewers said that the pipeline we used to generate the benchmark was also a research contribution, even though it&#8217;s the minimum viable scaffold: a loop in serial.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kogt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kogt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 424w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 848w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 1272w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kogt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png" width="1456" height="531" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:531,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kogt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 424w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 848w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 1272w, https://substackcdn.com/image/fetch/$s_!Kogt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b437cd0-64ba-4e83-903b-d0a4e2f25f2b_1600x584.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Say hi In Ottawa on ICSE workshop day! Special shoutout to <a href="https://gasstationmanager.github.io/ai/2025/02/18/fvapps.html">Gas Station Manager for citing us</a> like a week after we went public in their hallucination work. </p><h2>Is this a safety paper?</h2><p>I'm working on making sure we get high quality critical systems software out of early AGI. Hardened infrastructure buys us a lot in the slightly crazy story of "self-exfiltrated model attacks the power grid", but buys us even more in less crazy stories about all the software modules adjacent to AGI having vulnerabilities rapidly patched at crunchtime.</p><p>While I&#8217;m not super interested in measurement, I&#8217;m excited about this line of work as a <em>synthetic data pipeline</em>. It may take galaxy brained prompting and galaxy brained scaffolds to ship proofs now (if they&#8217;re elicitable at all), but by solving FVAPPS you generate synthetic data which you can put to use in finetune jobs, leading to models way better at proving, hopefully with minimal scaffolding and prompting skills. We shipped our paper to arxiv twitter the same week <a href="https://goedel-lm.github.io/">Goedel-LM</a> shipped, and what they do is a kind of &#8220;finetune job in-the-loop&#8221; along these lines (i.e. using Lean as a ground truth signal).</p><h1>R1 happened</h1><p>With finetuning this cheap, lots of opportunities are afoot. That&#8217;s it, that&#8217;s the post. Especially of interest is that API terms of service aren&#8217;t acceptable for some critical systems vendors for compliance reasons, so running a top performing model on compute you control opens up many possibilities for them in particular.</p><h1>News in automated mathematics</h1><h2>AlphaGeometry finally dropped a paper</h2><p>For a while Google DeepMind&#8217;s results in mathematics just had a vague blog post. <a href="https://arxiv.org/abs/2502.03544">It&#8217;s great there are finally details in a paper</a>, but I haven&#8217;t read it yet.</p><h2>New SOTA on MiniF2F</h2><p>https://goedel-lm.github.io/</p><h1>The topological debate framework</h1><p>I had a chance to sit down with the author at EAG. This is an exciting theory of the <em>incentives</em> for AIs to have good world models. These are the kind of protocols that could help us <strong>close the spec-world gap</strong>, which is going to be critical if GSAI does anything at all against arbitrary superintelligences.</p><blockquote><p>Let's say that you're working on a new airplane and someone hands you a potential design. The wings look flimsy to you and you're concerned that they might snap off in flight. You want to know whether the wings will hold up before you spend money building a prototype. You have access to some 3D mechanical modeling software that you trust. This software can simulate the whole airplane at any positive resolution, whether it be 1 meter or 1 centimeter or 1 nanometer.</p><p>Ideally you would like to run the simulation at a resolution of 0 meters. Unfortunately that's not possible. What can you do instead? Well, you can note that all sufficiently small resolutions should result in the same conclusion. If they didn't then the whole idea of the simulations approximating reality would break down. You declare that if all sufficiently small resolutions show the wings snapping then the real wings will snap and if all sufficiently small resolutions show the wings to be safe then the real wings will be safe.</p><p>How small is "sufficiently small?" A priori you don't know. You could pick a size that feels sufficient, run a few tests to make sure the answer seems reasonable, and be done. Alternatively, you could use the two computationally unbounded AI agents with known utility functions that you have access to.</p></blockquote><p>Debate is a little bit of a misnomer, but close enough.</p><p><a href="https://www.lesswrong.com/posts/jCeRXgog38zRCci4K/topological-debate-framework">https://www.lesswrong.com/posts/jCeRXgog38zRCci4K/topological-debate-framework</a></p><h1>Atlas hiring dev for spec engineering GUI product</h1><p>If we can kick the proof down to the AI and rely on the typechecker, we still want a human in the loop for writing the specification</p><blockquote><p>We are developing an AI-assisted IDE for formal specification.</p><p>Autoformalization is approaching usefulness but there remains a gap for how humans establish confidence in autoformalized specs and discover issues. We're researching tools and methods to enable engineers to close this gap in real-world assurance applications. We have been prototyping and user-testing a Spec IDE and are looking to add a team member to take over primary responsibility for machine learning engineering.</p><p>This job involves developing a machine learning pipeline that powers mechanized spec engineering and review. ML plays multiple roles in the spec IDE: (1) aiding user understanding and navigation by labeling concepts within and across both mechanized and pen-and-paper spec documents, (2) detecting possible issues in the mechanization, and (3) powering a conversational assistant for users to navigate or edit specs.</p><p>Autoformalization is not itself the focus of this project, but there's an opportunity to explore autoformalization if desired. This job involves collaborating with a small remote team that brings a user-centered, product discovery mindset to this research; as such, this job also involves learning from user research, and has opportunities to run user studies if desired.</p><p>We're looking for a 20 hrs/wk contract through approximately September, but there's room for flexibility. Please let us know your contracting rate ($/hr) if you have a standard rate, else reach out to chat with us; cost will be one of our selection factors.</p><p>Contact join@atlascomputing.org with interest.</p></blockquote><p>They want someone with 2 years of surface area on proof assistants like Lean and Coq and experience with ML R&amp;D tooling.</p><h1>Safeguarded AI TA1.2 and TA1.3 funding call</h1><p><a href="https://vimeo.com/1054492081/8f3ceee33f">Video</a>, <a href="https://www.aria.org.uk/media/jxrhq22y/safeguarded-ai-ta12-13-programme-solicitation-call.pdf">PDF</a>. Davidad is back with a new funding call.</p><p>Technical Area 1.1 is underway and consists of category theorists and probabilistic semantics experts <em>designing</em> a new stack for world models, proof certificates, and specifications. It&#8217;s the conceptual problems at the foundation of Safeguarded AI. Upon successful solutions to those problems, it becomes a fairly normal software project: a <em>backend</em> and a <em>frontend</em>.</p><p>TA1.2 is a backend. This involves structure-aware version control (<a href="https://gsai.substack.com/i/155386915/structure-aware-version-control-via-observational-bridge-types-hiring">previous edition of the newsletter</a>) for eliciting world models and specs, other data structures topics for keeping track of probabilistic proof certs, and an appropriate theory of databases.</p><p>TA1.3 is a frontend. Many important stress points of Safeguarded AI require human in the loop, so we want a killer user experience. Imagine being a power plant domain expert in the near future, when AIs can do anything, but they need to <strong>elicit</strong> specs and world models from you. Hopefully it&#8217;s not too painful a process!</p><p>Both projects are investing 12 months into requirements engineering with minimal prototyping, and the new org that TA2 is planning to incubate around the end of this year will be partially shaped by those requirements.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ckpU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ckpU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 424w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 848w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 1272w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ckpU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png" width="1456" height="815" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:815,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ckpU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 424w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 848w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 1272w, https://substackcdn.com/image/fetch/$s_!ckpU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d1c1e8-e9b3-446e-97de-f157583d3b02_1600x896.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I think the microeconomics of this ambition level could be reasoned about (scrutinizing the reference class of software projects roughly this big, counting up roughly how much money was spent on them, fermstimating if this is realistic, etc.). But it&#8217;s out of scope for this newsletter.</p><p>The deadline is April 8th. It&#8217;s secretly the 9th but it&#8217;s a good habit to subtract one day when you put it in your calendar, especially since it doesn't close on an anywhere-on-earth midnight.</p>]]></content:encoded></item><item><title><![CDATA[November-December 2024 Progress in Guaranteed Safe AI]]></title><description><![CDATA[You missed my hot takes on FrontierMath, ho ho ho]]></description><link>https://newsletter.for-all.dev/p/november-december-2024-progress-in</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/november-december-2024-progress-in</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Wed, 22 Jan 2025 01:12:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!F2-B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!F2-B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!F2-B!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 424w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 848w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!F2-B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg" width="751" height="499" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:499,&quot;width&quot;:751,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!F2-B!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 424w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 848w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!F2-B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6cb6627-4330-48aa-b15f-91385de9547f_751x499.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Sorry for the radio silence last month. It was slow and I didn&#8217;t come across things I wanted to write about, to be expected with holidays coming up.</p><p>There are no benefits of paying, except you get a cut of my hard earned shapley points, and apparently some disappointment when I miss a month.</p><p>If you're just joining us,<a href="https://arxiv.org/abs/2405.06624"> background on GSAI here</a>.</p><p>There should be some DeepSeek/r1 coverage in the next newsletter (I think tanking the cost of finetuning is, while dual use, good news for us in GSAI). I&#8217;d rather get this one out the door though.</p><h1>FrontierMath (hiring)</h1><p>As I&#8217;m always banging on about, progress in math automation is a leading indicator for progress in software engineering automation by the curry-howard correspondence. That&#8217;s a little cheeky, but last time I said it in meatspace I got solemn nods as if it wasn&#8217;t even cheeky at all. Or maybe they were solemnly nodding at something else. To be explicit, mathematical reasoning abilities and software engineering skills (including formal proof engineering) appear to be at similar capability levels. Furthermore, through platforms like Lean, mathematical ability can enhance software development capabilities. This connection should make it easier to implement rigorous quality assurance processes, such as obtaining formal proof certificates, which are typically costly.</p><blockquote><p>We introduce FrontierMath, a benchmark of hundreds of original, exceptionally challenging mathematics problems crafted and vetted by expert mathematicians. The questions cover most major branches of modern mathematics -- from computationally intensive problems in number theory and real analysis to abstract questions in algebraic geometry and category theory. Solving a typical problem requires multiple hours of effort from a researcher in the relevant branch of mathematics, and for the upper end questions, multiple days. FrontierMath uses new, unpublished problems and automated verification to reliably evaluate models while minimizing risk of data contamination. Current state-of-the-art AI models solve under 2% of problems, revealing a vast gap between AI capabilities and the prowess of the mathematical community. As AI systems advance toward expert-level mathematical abilities, FrontierMath offers a rigorous testbed that quantifies their progress.</p></blockquote><p>The story here of course is that by the time the newsletter covered it, <a href="http://12daysofopenai.com/#day12">Open AI hit 25%</a>. &#8220;Mathematicians assemble a benchmark they can barely solve themselves at SOTA 2%, forecasters think it&#8217;ll be unsaturated for multiple years, a month later Open AI hits 25%&#8221; is a fun little anecdote for Christmas with the relatives, for a funny sense of fun. But, <a href="https://x.com/CloutUnmatched/status/1881538415398891882?t=5pjvj8oINbzG9z8_pFoSIA&amp;s=19">points out an astute reader</a>, since we have typical Open AI science communication (read: extremely disappointing science communication) here: what sort of inference compute was thrown at it? <em>At what cost?</em> A model that can win with the most naive prompt is more impressive than a model that only wins with gal-brained and/or expensive (in tokens) feedback schemes.</p><p>So anyways, the story here of course is that after I drafted the previous paragraph, people noticed that Epoch sneakily added their funding disclosure to a new draft of the paper a month after all the hoopla had started to die down. Spoiler alert, the funder was, lol, Open AI. I&#8217;ll refer you to <a href="https://www.lesswrong.com/posts/8ZgLYwBmB3vLavjKE/some-lessons-from-the-openai-frontiermath-debacle">LessWrong user 7vik&#8217;s summary of events</a>, and not get too far into it.</p><p>But I have <em>feelings</em>. What if the externality of mathematical automation and tooling in the form of progress in guaranteed safe AI doesn&#8217;t pan out? Then Open AI gets it&#8217;s &#8220;put terry tao out of a job&#8221; boyscout badge for all the employee laptops and the safety community gets nothing in return! Come on guys, why is this even in Epoch&#8217;s mandate to begin with? Makes me sad.</p><h2>See also</h2><p><a href="https://arxiv.org/abs/2412.15184">Review paper &#8220;The suite of datasets commonly used to train and evaluate the mathematical capabilities of AI-based mathematical copilots (primarily large language models) exhibit several shortcomings&#8221;</a>, <a href="https://www.reddit.com/r/math/comments/1h6rwls/im_developing_frontiermath_an_advanced_math/">Elliot&#8217;s r/math AMA</a>, <a href="https://xenaproject.wordpress.com/2024/12/22/can-ai-do-maths-yet-thoughts-from-a-mathematician/">Kevin Buzzard on FrontierMath</a> (written after Open AI results).</p><h2>Opportunity</h2><p>FrontierMath begins in three tiers of difficulty (25% undergrad or IMO level, 50% grad student, 25% early professor-track researcher), but they&#8217;re currently expanding to a fourth tier of even harder problems, and they also want some formalization in Lean, which you email elliot at epochai.org about if you&#8217;d like to partake.</p><h1>ARIA call: Mathematics for Safe AI (funding)</h1><p><a href="https://www.aria.org.uk/opportunity-spaces/mathematics-for-safe-ai/mathematics-for-safe-ai">This round</a> is affiliated with the Safeguarded AI programme, but is outside of the standard technical areas breakdown.</p><h2>See also</h2><p>Renaissance Philanthropy&#8217;s <a href="https://renaissancephilanthropy.org/initiatives/ai-for-math-fund/">AI for Math fund</a> (unclear how much this will help with GSAI efforts).</p><h1>Structure-aware version control via observational bridge types (hiring)</h1><p><a href="https://topos.institute/blog/2024-11-13-structure-aware-version-control-via-observational-bridge-types/#fn6">David Jaz Myers writes for the topos blog</a> about structure-aware version control.</p><p>This is really exciting, because git diffs as we know them would lead to problems if you were scaling world models (in a davidad-style approach) or specifications (in any GSAI approach including davidad) that were being collaborated on with many humans and AIs.</p><blockquote><p>But suppose that you and I are editing a<a href="https://en.wikipedia.org/wiki/Comma-separated_values"> </a>csv, and you add a row to the bottom and I add a column to the end. Git would see your change as a single line diff, whereas my change (adding a column) is a change to every line; these conflict in the new line you added. But from what we know about the structure of csv, your change should really be a one-row change, mine a one-column change, and the conflict should only occur in one cell.</p></blockquote><p>The author proposes something called <em>observational bridge types</em> (from the up and coming proof assistant Narya) to form the foundation of structure-aware version control. Using these, we can say that, for filetype F and files f1 f2 : F, a <strong>diffing algorithm</strong> is none other than any inhabitant of the type Diff F f1 f2 (where Diff : (A : Type) -&gt; A -&gt; A -&gt; Type is created using something called <em>logical relations</em>, or an inductive definition depending on a type). Then, <strong>conflicts</strong><em><strong> </strong></em>are defined as a pair (d1, d2) : Diff F f1 f2 x Diff F f1 f3 so that <strong>merges</strong> may be defined as another pair (d3, d4) : Diff F f2 f4 x Diff F f3 f4. That much is roughly consistent with git, provided that you assume F (the filetype) is always &#8220;list[list[char]]&#8221; or some notion of raw text. It&#8217;d be great to not have to assume that, even just for the csv example, to say nothing of more complicated structures in probabilistic semantics or complex/chaotic systems. A system that can also reason about diffs between the types themselves (since Diff Type sigma tau is perfectly reasonable here) is also desirable.</p><h2>See also</h2><p>More on this <a href="https://youtu.be/nN6p9u-_Tro?si=oNMs5Bspjl8GfxUz">from Owen at Topos at last year&#8217;s Safe By Design meeting at FAR Labs</a>. More on this <a href="https://groups.google.com/g/guaranteed-safe-ai/c/eqnQwwKsz0U/m/4I8gxRGLBgAJ">from the GSAI google group a while ago</a>. <a href="https://difftastic.wilfred.me.uk/tree_diffing.html">Roundup of tree diff tools</a>.</p><h2>Opportunity</h2><p>To work with David Jaz on this for the world modeling part of Safeguarded AI in Oxford, <a href="https://topos.institute/files/jobs/oxford-2024-masd-postdoc.pdf">apply here</a>.</p><h1>Squiggle AI</h1><p>In the davidad and somewhat Bengio regions of the GSAI space, there&#8217;s a huge emphasis on world models&#8211; computational descriptions of what is. Squiggle is a programming language where the terms are distributions, and all monte carlo boilerplate is hidden from the user. This makes Squiggle a candidate &#8220;standard&#8221; for belief specifications, <a href="https://www.lesswrong.com/posts/qg4yy4je3sFm7Z49L/what-is-estimational-programming-squiggle-in-context">as I&#8217;ve written about elsewhere</a>.</p><p>The first problem you&#8217;d run into is that the Users Will Not Just. The users will not just learn a programming language to specify beliefs when they barely want to specify their beliefs in the first place. That&#8217;s why it&#8217;s good news QURI shipped an AI product, so the interface is closer to natural language while the substrate/backend is squiggle. The LLM Will Just. The LLM will just learn to program in squiggle, so you don&#8217;t have to.</p><p>It&#8217;s ready for you to use it at <a href="http://squigglehub.org/ai">squigglehub.org/ai</a> (if authenticated), and <a href="https://forum.effectivealtruism.org/posts/jJ4pn3qvBopkEvGXb/introducing-squiggle-ai">you can read QURI&#8217;s press release on the EA Forum and elsewhere</a>.</p>]]></content:encoded></item><item><title><![CDATA[October 2024 Progress in Guaranteed Safe AI]]></title><description><![CDATA[Safeguarded AI TA 1.4 funding call, a way of verifying quantitative properties of program pairs, and previews of the recent ICLR submissions.]]></description><link>https://newsletter.for-all.dev/p/october-2024-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/october-2024-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Mon, 28 Oct 2024 23:16:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e6f92c2-2ffc-4b0b-ad92-596137aa1263_576x576.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There are no benefits of subscribing, except you get a cut of my hard earned shapley points.&nbsp;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.for-all.dev/subscribe?"><span>Subscribe now</span></a></p><p>As always, do reach out to inform me how the newsletter could be better, unless you want less biased and wrong quick takes. The biased and wrong quick takes are <em>mandatory</em>. Do you like inlining abstracts, or should I just expect people to click when they want to read the abstract? This month we&#8217;re inlining all of them.&nbsp;</p><p>If you're just joining us, <a href="https://arxiv.org/abs/2405.06624">background on GSAI here</a>.</p><p>I attended an ARIA meeting this month, so most of the newsletter is papers I found there that happen to be recent-ish.</p><p>A little birdie told me there&#8217;s a meeting in Berkeley at the beginning of December&#8211; reach out for more info!</p><h1><a href="https://leanprover.zulipchat.com/#narrow/channel/219941-Machine-Learning-for-Theorem-Proving/topic/ICLR.202025.20Submissions/near/475186957">Big list of openreview links for ICLR 2025 papers related to scaled interactive theorem proving</a>&nbsp;</h1><p>From the Lean Zulip. There&#8217;s probably more from ICLR I&#8217;d like to cover in the newsletter, but I haven&#8217;t parsed the openreview firehose personally. I&#8217;m freeloading off Lean Zulip users. </p><p>One thing that jumped out at me is in <a href="https://leanprover.zulipchat.com/#narrow/channel/219941-Machine-Learning-for-Theorem-Proving/topic/LeanAgent">a thread about a polarizing/unpopular paper</a>, the discussion of what programming practices would lead to ML papers having an easier time at making reasonable claims about the nontriviality of discovered proofs and broadly preventing contamination at scale. Terence Tao discusses placeholder design patterns, suggests a proof_wanted command to replace a theorem &#8230; : 0 = 1 := sorry, and discusses downsides with the practice of sorrying out 0=1 to do placeholding (spoiler: an AI can exploit the proof of a false proposition to misbehave). This is an important point: any GSAI stack will have collaboration between proof engineers of different power levels, and lots of collaboration in proof engineering is driven by very careful bookkeeping of placeholders (saying what we need, what we have, and what would be nice to have).&nbsp;</p><h1><a href="https://arxiv.org/abs/2407.17127">A quantitative probabilistic relational hoare logic</a></h1><p><a href="https://calendar.app.google/2a11YNXUFwzHbT3TA">November&#8217;s GSAI paper club</a>.&nbsp;</p><p><strong>Hoare logic</strong> is when you have precondition <em>P : Memory -&gt; Prop</em> and postcondition <em>Q : Memory -&gt; Prop</em>, and some imperative program <em>c</em> such that forall memories <em>m </em>and <em>n</em>, if <em>Pm</em> and executing <em>c</em> sends <em>m</em> to <em>n</em>, then you can prove <em>Qn</em>, written {P}c{Q}. Sometimes, you want to reason about relations between programs. <strong>Relational hoare logic</strong><em><strong> </strong></em><strong>(RHL) </strong>is when preconditions and postconditions are relations on memory states and you have two programs, so {P} c&#8321; ~ c&#8322; {Q} means that forall memories <em>m1 m2 n1 n2</em>, if <em>P m1 m2</em> and <em>exec c1 m1 = n1</em> and <em>exec c2 m2 = n2</em>, then <em>Q n1 n2</em>.&nbsp;</p><p>You might like to write a proof of a RHL quadruple for a cryptographic primitive, requiring you to reason about independence of two PRNGs. Independence is one of the things you don&#8217;t get in RHL, because RHL assumes deterministic programs. <strong>Probabilistic relational hoare logic (pRHL)</strong> looks again like {P} c1 ~ c2 {Q}, but this time Q is lifted to a relation between distributions over memory, reflecting the possible outcomes of programs involving probabilistic choice or random sampling. So precondition P is of type Memory -&gt; Memory -&gt; Prop, postcondition Q is of type &#916;Memory -&gt; &#916;Memory -&gt; Prop, and the pRHL quadruple unpacks to forall <em>m1 m2 : Memory, </em>forall<em> N1 N2 : &#916;Memory,</em> if <em>P m1 m2</em> and <em>exec c1 m1 ~ N1 </em>and exec c2 m2 ~ N2, then <em>Q N1 N2</em>.&nbsp;</p><blockquote><p>We introduce eRHL, a program logic for reasoning about relational expectation properties of pairs of probabilistic programs. eRHL is quantitative, i.e., its pre- and post-conditions take values in the extended non-negative reals. Thanks to its quantitative assertions, eRHL overcomes randomness alignment restrictions from prior logics, including PRHL, a popular relational program logic used to reason about security of cryptographic constructions, and apRHL, a variant of PRHL for differential privacy. As a result, eRHL is the first relational probabilistic program logic to be supported by non-trivial soundness and completeness results for all almost surely terminating programs. We show that eRHL is sound and complete with respect to program equivalence, statistical distance, and differential privacy. We also show that every PRHL judgment is valid iff it is provable in eRHL. We showcase the practical benefits of eRHL with examples that are beyond reach of PRHL and apRHL.</p></blockquote><p>pRHL is not <em>expectational</em>. Since preconditions and postconditions land in Prop, we can&#8217;t <em>value</em> preconditions and postconditions as quantities. Last july, the authors introduced a generalization of pRHL called eRHL (e for expectation), where now preexpectations and postexpectations are of type Memory -&gt; Memory -&gt; NNReal, and the rest follows. eRHL is special because they can prove completeness theorems for almost surely terminating (Pr(halt) = 1) programs that pRHL can&#8217;t, along with some fancy desiderata like a statistical difference &#8660; eRHL quadruple correspondence and differential privacy.&nbsp;</p><p>If you&#8217;re wondering why you just read all that, here&#8217;s the juice: often in GSAI position papers there&#8217;ll be some reference to expectations that capture &#8220;harm&#8221; or &#8220;safety&#8221;. Preexpectations and postexpectations with respect to particular pairs of programs could be a great way to cash this out, cuz we could look at programs as interventions and simulate RCTs (labeling one program control and one treatment) in our world modeling stack. When it comes to harm and safety, Prop and bool are definitely not rich enough.&nbsp;&nbsp;</p><h1><a href="https://lu.ma/4ylbvs75">Bengio talk coming up&nbsp;</a></h1><p>November 14, 18:00-19:00 UTC</p><blockquote><p>Description: Could there be safety advantages to the training of a Bayesian oracle that is trained to only do that job, i.e., estimate P(answer | question, data)? What are the scenarios in which such an AI could cause catastrophic harm? Can we even use such an oracle as the intelligence engine of an agent, e.g., by sampling actions that help to achieve goals? What can go wrong even if we assume that we have a perfect prediction of the Bayesian posterior, e.g., if the true explanatory theory is a minority voice in the Bayesian posterior regarding harm prediction? If such an oracle is estimated by a neural network with amortized inference, what could go wrong? Could the implicit optimization used to train the estimated posterior create loopholes with an optimistic bias regarding harm? Could we also use such a Bayesian oracle to obtain conservative risk estimates, i.e., bounds on the probability of harm, that can mitigate the imperfections in such an agent?</p></blockquote><h1><a href="https://arxiv.org/abs/2409.13082">AutoVerus</a></h1><p>Verus is a framework for <em>annotating</em> rust code with preconditions (&#8220;requires&#8221;) and postconditions (&#8220;ensures&#8221;) and kicks static verification down to SMT. Here&#8217;s an example program for multiplying a number by 8:</p><p>fn octuple(x1: i8) -&gt; (x8: i8)</p><p>&nbsp;&nbsp;&nbsp;&nbsp;requires</p><p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;-16 &lt;= x1,</p><p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;x1 &lt; 16,</p><p>&nbsp;&nbsp;&nbsp;&nbsp;ensures</p><p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;x8 == 8 * x1,</p><p>{</p><p>&nbsp;&nbsp;&nbsp;&nbsp;let x2 = x1 + x1;</p><p>&nbsp;&nbsp;&nbsp;&nbsp;let x4 = x2 + x2;</p><p>&nbsp;&nbsp;&nbsp;&nbsp;x4 + x4</p><p>}</p><p>Without the &#8220;require&#8221;, we may overflow the 8 bit integer. The &#8220;ensure&#8221; is a functional specification of what it means for the function to be correct. The verus paper is <a href="https://arxiv.org/abs/2303.05491">here</a> and the verus docs are <a href="https://verus-lang.github.io/verus/guide/overview.html">here</a>.&nbsp;</p><blockquote><p>Generative AI has shown its values for many software engineering tasks. Still in its infancy, large language model (LLM)-based proof generation lags behind LLM-based code generation. In this paper, we present AutoVerus. AutoVerus uses LLM to automatically generate correctness proof for Rust code. AutoVerus is designed to match the unique features of Verus, a verification tool that can prove the correctness of Rust code using proofs and specifications also written in Rust. AutoVerus consists of a network of LLM agents that are crafted and orchestrated to mimic human experts' three phases of proof construction: preliminary proof generation, proof refinement guided by generic tips, and proof debugging guided by verification errors. To thoroughly evaluate AutoVerus and help foster future research in this direction, we have built a benchmark suite of 150 non-trivial proof tasks, based on existing code-generation benchmarks and verification benchmarks. Our evaluation shows that AutoVerus can automatically generate correct proof for more than 90% of them, with more than half of them tackled in less than 30 seconds or 3 LLM calls.&nbsp;</p></blockquote><p>The autoverus paper is an example in the genre of using verification tools as the ground truth in a machine learning process. A proof system is a great data labeler. GSAI is happy when this basic viewpoint makes data quality more scalable than other forms of ground truth across machine learning.&nbsp;</p><h1><a href="https://arxiv.org/abs/2408.15572">Sufficient and Necessary Barrier-like Conditions for Safety and Reach-avoid Verification of Stochastic Discrete-time Systems</a></h1><p>A reach-avoid specification is something of the form &#8220;eventually P and always not Q&#8221;. Barrier certs are a way of proving that a system starting in the safe region stays in the safe region without directly solving the whole system of differential equations.&nbsp;</p><blockquote><p>In this paper, we examine sufficient and necessary barrier-like conditions for the safety verification and reach-avoid verification of stochastic discrete-time systems. Safety verification aims to certify the satisfaction of the safety property, which stipulates that the probability of the system, starting from a specified initial state, remaining within a safe set is greater than or equal to a specified lower bound. A sufficient and necessary barrier-like condition is formulated for safety verification. In contrast, reach-avoid verification extends beyond safety to include reachability, seeking to certify the satisfaction of the reach-avoid property. It requires that the probability of the system, starting from a specified initial state, reaching a target set eventually while remaining within a safe set until the first hit of the target, is greater than or equal to a specified lower bound. Two sufficient and necessary barrier-like conditions are formulated under certain assumptions. These conditions are derived via relaxing Bellman equations.&nbsp;</p></blockquote><h1><a href="https://arxiv.org/abs/2409.10018">Compositional Design of Safety Controllers for Large-scale Stochastic Hybrid Systems</a></h1><p>What is the problem? When is a solution &#8220;compositional&#8221;?&nbsp;</p><p>The problem is stochastic hybrid systems (SHS), which are very heterogeneous: dynamics (i.e. of the kind described by differential equations) that might be discrete time in some parts and continuous time in others, and discrete state in some parts but continuous state in others. Specifically, the problem is synthesizing guaranteed controllers for these systems (some actuator that intervenes to enact a predicted output). A solution to a problem is compositional when it allows you to piece together solutions to bigger problems out of solutions to smaller problems.&nbsp;</p><blockquote><p>In this work, we propose a compositional scheme based on small-gain reasoning for the safety controller synthesis of interconnected stochastic hybrid systems with both continuous evolutions and instantaneous jumps. In our proposed setting, we first offer an augmented scheme to represent each stochastic hybrid subsystem with continuous and discrete evolutions in a unified framework, ensuring that the state trajectories match those of the original hybrid systems. We then introduce the concept of augmented control sub-barrier certificates (A-CSBC) for each subsystem, which allows the construction of augmented control barrier certificates (A-CBC) for interconnected systems and their safety controllers under small-gain compositional conditions. We eventually leverage the constructed A-CBC and quantify a guaranteed probabilistic bound across the safety of the interconnected system. While the computational complexity of designing a barrier certificate and its safety controller grows polynomially with network dimension using sum-of-squares (SOS) optimization program, our compositional approach significantly reduces it to a linear scale with respect to the number of subsystems. We verify the efficacy of our proposed approach over an interconnected stochastic hybrid system composed of 1000 nonlinear subsystems.</p></blockquote><h1><a href="https://arxiv.org/abs/2406.08467">DafnyBench (back in june)</a></h1><p>The Beneficial AI Foundation and others shipped a benchmark for Dafny coding.&nbsp;</p><blockquote><p>We introduce DafnyBench, the largest benchmark of its kind for training and evaluating machine learning systems for formal software verification. We test the ability of LLMs such as GPT-4 and Claude 3 to auto-generate enough hints for the Dafny formal verification engine to successfully verify over 750 programs with about 53,000 lines of code. The best model and prompting scheme achieved 68% success rate, and we quantify how this rate improves when retrying with error message feedback and how it deteriorates with the amount of required code and hints. We hope that DafnyBench will enable rapid improvements from this baseline as LLMs and verification techniques grow in quality.&nbsp;</p></blockquote><p>There&#8217;s also <a href="https://github.com/Beneficial-AI-Foundation/dafny-autopilot">this VSCode extension</a> complimenting the paper.&nbsp;</p><h1><a href="https://t.co/JgGHjmduvJ">Safeguarded AI TA1.4 funding call</a></h1><p>Economists and voting theorists are being summoned to help us reason about specification generating processes under multiple stakeholders. <a href="https://t.co/U04v2lH2d0">In video form!</a></p><p>We already knew that ARIA is being wildly ambitious. I&#8217;d be happy if we could get <em>any specs at all</em> on the new spec/modeling/cert stack that ARIA is building to be checked against prod code; it seems a lot harder to make them also socially/democratically desirable. Worried a little about scope creep with these guys, but what else is new.&nbsp;</p>]]></content:encoded></item><item><title><![CDATA[September 2024 Progress in Guaranteed Safe AI]]></title><description><![CDATA[This month we have elicitation of GPT-o1&#8217;s Lean4 skills, a worked example in self driving cars, and an interactive proof approach to correctness of learned components.]]></description><link>https://newsletter.for-all.dev/p/september-2024-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/september-2024-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Thu, 26 Sep 2024 20:13:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/608d616c-3efe-4c05-a2a2-ded698a186d0_508x284.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There are no benefits with paid subscription, except you get a cut of my hard earned shapley points.&nbsp;</p><p>As always, do reach out to inform me how the newsletter could be better, unless you want less biased and wrong quick takes. The biased and wrong quick takes are <em>mandatory</em>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Progress in Guaranteed Safe AI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>If you're just joining us, <a href="https://arxiv.org/abs/2405.06624">background on GSAI here</a>.&nbsp;&nbsp;</p><h1>Terry Tao&#8217;s mathstodon comment</h1><p><a href="https://mathstodon.xyz/@tao/113132502735585408">&#8220;I have played a little bit with OpenAI&#8217;s new iteration of GPT&#8221;, he begins</a>.&nbsp;</p><p>Let's get everyone up to speed on one of Tao's key insights: the bottlenecks to casually superb Lean4 performance are embarrassingly non-profound, chat models getting tripped up by import paths changing across Lean versions and so on. You can only choke a model by having too much code velocity in the training data for so long, but eventually that'll fall.&nbsp;</p><p>That may not be a great crux for Tao's use case. I think he wants to focus on the reasoning bottlenecks, and is less concerned about coding bottlenecks given reasoning. Perhaps his interest in the Lean parts is mostly Voevedsky style: what do we want? Less homework. How do we want it? With greater assurance than more homework.&nbsp;</p><p>You should sit and think about if you think attaining a reasoning level is easier or harder than turning that reasoning level into Lean code once attained. I sat and thought about it for 5 minutes, and I'm still not ready to make any bets. Yes, the lesson about the embarrassingly non-profound bottlenecks suggest a reasoning level is harder than translating that reasoning level into code, but on the other hand, dependent types are a really finicky way to program.&nbsp;</p><p>Tao reports that we've leapt from incompetent grad student to mediocre grad student, or openai has anyway. I do not know the conversion factor between units of pure maths grad student to units of GSAI-grade software developer. I'll take 1 to 1 as my prior. How about yours? Is one pure maths grad student worth of software engineering enough to accomplish your goals, if you can make copies and replace pizza/coffee with electricity?&nbsp;</p><h1>David Manheim&#8217;s post</h1><p>I liked <a href="https://www.lesswrong.com/posts/3P8WBwLyfPBEkbG3c/proveably-safe-self-driving-cars">David Manheim&#8217;s post on self driving cars as a worked example</a>. Like we <a href="https://gsai.substack.com/i/148170804/a-wager-you-say">saw last month</a>, worked examples lead to identifying cruxes clearer and faster. In this case, the commenters mostly objected about unstated assumptions, and Manheim would go on to edit more assumptions (even in the title of the post!).&nbsp;</p><p>Tdietterich:&nbsp;</p><blockquote><p>Perhaps we should rename the "provable safety" area as "provable safety modulo assumptions" area and be very explicit about our assumptions. We can then measure progress by the extent to which we can shrink those assumptions.</p></blockquote><p>Jacobjacob:</p><blockquote><p>Very much agree. I gave some feedback along those lines as the term was coined; and am sad it didn't catch on. But of course "provable safety modulo assumptions" isn't very short and catchy...</p><p>I do like the word "guarantee" as a substitute. We can talk of formal guarantees, but also of a store guaranteeing that an item you buy will meet a certain standard. So it's connotations are nicely in the direction of proof but without, as it were, "proving too much" :)</p></blockquote><p>I agree with the sentiment in the comments that provability in the sense of system safety is provability up to <em>which</em> assumptions, and hope to see people be more clear about this.&nbsp;</p><h1>Models that prove their own correctness&nbsp;</h1><p><a href="https://arxiv.org/abs/2405.15722">This was june but I just found it</a>. We&#8217;re reading it for next month&#8217;s <a href="https://calendar.app.google/ZpELWosURvz1dYgn6">GSAI paper club, which is convening 2024-10-17 at 10a pacific</a>. Here, the word &#8220;proof&#8221; is in the context of &#8220;interactive proof&#8221;, or some multi-step game consisting of a prover and a verifier popular in computational complexity theory and cryptography.&nbsp;</p><blockquote><p>How can we trust the correctness of a learned model on a particular input of interest? Model accuracy is typically measured *on average* over a distribution of inputs, giving no guarantee for any fixed input. This paper proposes a theoretically-founded solution to this problem: to train *Self-Proving models* that prove the correctness of their output to a verification algorithm V via an Interactive Proof. Self-Proving models satisfy that, with high probability over a random input, the model generates a correct output *and* successfully proves its correctness to V. The *soundness* property of V guarantees that, for *every* input, no model can convince V of the correctness of an incorrect output. Thus, a Self-Proving model proves correctness of most of its outputs, while *all* incorrect outputs (of any model) are detected by V. We devise a generic method for learning Self-Proving models, and we prove convergence bounds under certain assumptions. The theoretical framework and results are complemented by experiments on an arithmetic capability: computing the greatest common divisor (GCD) of two integers. Our learning method is used to train a Self-Proving transformer that computes the GCD *and* proves the correctness of its answer.</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dkPX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dkPX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 424w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 848w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 1272w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dkPX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png" width="508" height="284" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a119b208-6638-4738-88be-805a77436c1e_508x284.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:284,&quot;width&quot;:508,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dkPX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 424w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 848w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 1272w, https://substackcdn.com/image/fetch/$s_!dkPX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa119b208-6638-4738-88be-805a77436c1e_508x284.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Some people in GSAI only want to gain assurance about code that AIs are writing, others want to gain assurance about the learned components themselves. If you&#8217;re in the latter camp, this proposal looks promising, but there&#8217;s a long road ahead to scale it to non-toy problems. </p><h1>Types Are Not Safe</h1><p>I came across <a href="https://youtu.be/QLXYL-Ln2PM?si=mFwKc9qjUkhU0FtU">this lambda conf talk</a> solely because I used to work for the speaker&#8211; I did not check whether other lambdaconf talks are worthy of endorsement.&nbsp;</p><p>I'm including it in the newsletter to make sure we get on the same page about what type systems can't do. Type systems check for a pretty narrow kind of correctness. No one has lifted color blind approval of UIs to types. No one has made a PDE solver where the typechecker knows if the initial conditions you loaded in send the plane into a hillside. Lattice based access control is kinda like lifting security to a type checker, but it doesn&#8217;t know which permission sets are more or less susceptible to insider, phishing, or bribery risks. This matters, because if you cash out &#8220;formal verification&#8221; at Lean, then you've just staked it all on types, and types are not safe. I think we need a broad portfolio and several different meanings of the word &#8220;certificate&#8221;, but I keep running into people who let their guard down and hail the coming age of formally verified code generation, and if they were careful they wouldn't be so optimistic.&nbsp;</p><p>I tend to say that type systems are for increasing the jurisdiction of compiletime knowledge. The runtime is where your adversary has home team advantage. But many correctness notions that are in principle liftable to a sufficiently expressive typechecker are not in practice worth the trouble. Now, this is due to ergonomic bottlenecks that an LLM could steamroll over, but then you're just shifting your error/attack surface from the spec-implementation gap to &#8220;where the h!ck did this spec come from?&#8221; and of course the ever present world-spec gap.</p><p>Another title for the talk could've been &#8220;Types Lie&#8221;, but that section of the talk relies a little on advanced language extensions specific to haskell that it isn't important you understand. Some of them, like the untypeability of the type of types, is way less bad in dependently typed languages. Getting confused because you wrote some type algebra that accidentally involves uninhabited types happened to me in Lean a couple weeks ago, though.&nbsp;</p><h1>Opportunities</h1><h2>Topos jobs&nbsp;</h2><p>Both of the <a href="https://topos.institute/contact">Oxford postdocs</a> are concerned with world modeling and the foundations of what eventually will be specification engineering for ARIA-style Safeguarded AI.&nbsp;</p><h2>UK&#8217;s AISI cites GSAI as an area they want to branch more into, are hiring</h2><p>The <a href="https://www.aisi.gov.uk/work/safety-cases-at-aisi">UK&#8217;s AI Safety Institute is building a team</a> for articulating positive &#8220;safety cases&#8221;, stories in which deploying models of the future is not a bad idea. They want technical talent to write these stories not just so that they&#8217;ll be calibrated but also so the team has the capacity in house to deliver on some of the empirical needs of those stories. They do plan to open an SF office over the next year, but right now you need to be onsite in London.&nbsp;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Progress in Guaranteed Safe AI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[August 2024 Progress in Guaranteed Safe AI]]></title><description><![CDATA[$10k on the line, RL from Proof Assistant Feedback, Aria&#8217;s TA3]]></description><link>https://newsletter.for-all.dev/p/august-2024-progress-in-guaranteed</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/august-2024-progress-in-guaranteed</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Tue, 27 Aug 2024 02:42:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!m7Qr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>A wager you say</h1><p>One proof of concept for the GSAI stack would be a well-understood mechanical engineering domain automated to the next level and certified to boot. How about locks? Needs a model of basic physics, terms in some logic for all the parts and how they compose, and some test harnesses that simulate an adversary. Can you design and manufacture a <em>provably unpickable</em> lock?&nbsp;</p><p><a href="https://www.lesswrong.com/posts/P8XcbnYi7ooB2KR2j/provably-safe-ai-worldview-and-projects?commentId=Ku3X4QDBSyZhrtxkM">Zac Hatfield-Dodds (of hypothesis/pytest and Anthropic, was offered and declined authorship on the GSAI position paper) challenged Ben Goldhaber to a bet after Ben coauthored a post with Steve Omohundro</a>. It seems to resolve in 2026 or 2027, the comment thread should get cleared up once Ben gets back from Burning Man. The arbiter is Raemon from LessWrong.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">to your inbox!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Zac says you can&#8217;t get a provably unpickable lock on this timeline. Zac gave (up to) 10:1 odds, so recall that the bet can be a positive expected value for Ben even if he thinks the event is most likely not going to happen.</p><p>For funsies, let&#8217;s map out one path of what has to happen for Zac to pay Ben $10k. This is not <em>the </em>canonical path, but it is <em>a</em> path:</p><ol><li><p>Physics to the relevant granularity (question: can human lockpicks leverage sub-newtownian issues?) is conceptually placed into type theory or some calculus. I tried a riemann integral in coq once (<em>way</em> once), so it occurs to me that you need to decide if you want just the functional models (perhaps without computation / with proof irrelevance) in your proof stack or if you want the actual numerical analysis support in there as well.</p></li><li><p>Good tooling, library support, etc. around that conceptual work (call it <em>mechlib</em>) to provide mechanical engineering primitives</p></li><li><p>A lock designing toolkit, depending on <em>mechlib</em>, is developed</p></li><li><p>Someone (e.g. a large language model) is really good at programming in the lock designing toolkit. They come up with a spec L.&nbsp;</p></li><li><p>You state the problem &#8220;forall t : trajectories through our physics simulation, if L(t) == open(L) then t == key(L)&#8221;</p></li><li><p><em>Then you get to write a nasty gazillion line Lean proof</em></p></li><li><p>Manufacture a lock (did I mention that the design toolkit has links to actual manufacturing stacks?)</p></li><li><p>Bring a bunch to DefCon 2027 and send another to the <a href="https://en.wikipedia.org/wiki/LockPickingLawyer">lockpicking lawyer</a></p></li><li><p>Everyone fails. Except Ben and the army of postdocs that $9,999 can buy.</p></li></ol><p>Looks like after the magnificent research engineering in steps 1 and 2, the rest is just showing off and justifying those two steps. Of course, in a world where we have steps 1 and 2 we have a great deal of transformative applications of formal modeling and verification just in reach, and we&#8217;ll need a PoC like locks to practice and concretize the workflow.&nbsp;</p><p>Cryptography applications tend to have a curse of requiring a lot of work after the security context, permission set, and other requirements are frozen in stone, which means that when the requirements change you have to start over and throw out a bunch of work (epistemic status: why do you think so many defi projects have more whitepapers than users?). The provably unpickable lock has 2 to 10 x that problem&#8211; get the granularity wrong in step one, most of your <em>mechlib</em> implementation won&#8217;t be salvageable. As the language model iterates on the spec L in step 5, the other language model has to iterate on the proof in step 6, because the new spec will break most of the proof.</p><p>Sorry I don&#8217;t know any mechanical engineering, Ben, otherwise I&#8217;d take some cracks at it. The idea of a logic such that its denotation is a bunch of mechanical engineering primitives seems interesting enough that my &#8220;if it was easy to do in less than a year someone would&#8217;ve, therefore there must be a moat&#8221; heuristic is tingling. Perhaps oddly, the quantum semantics <a href="https://arxiv.org/abs/quant-ph/0402130">folks</a> (<a href="https://www.researchgate.net/publication/380071047_Quantum_HoTT-Wiring_Weaving_Homotopy_Type_Theory_into_Quantum_Wave_Mechanics">or with HoTT!</a>) seem to have been productive, but I don&#8217;t know how much of that is translatable to mechanical engineering.</p><h1>Reinforcement learning from proof assistant feedback, and yet more monte carlo tree search</h1><p><a href="https://arxiv.org/abs/2408.08152">DeepSeek&#8217;s paper</a>&nbsp;</p><p>The steps are pretraining, supervised finetuning, RLPAF (reinforcement learning from proof assistant feedback), and MCTS (monte carlo tree search). RLPAF is not very rich: it&#8217;s a zero reward for any bug at all and a one for a happy typechecker. Glad they got that far with just that.&nbsp;</p><p><a href="http://deepseek.com">You can use the model at deepseek.com</a>.</p><h1>Harmonic ships their migration of miniF2F to Lean 4, gets 90% on it, is hiring</h1><p><a href="https://www.harmonic.fun/news">From their &#8220;one month in&#8221; newsletter</a>. &#8220;Aristotle&#8221;, which has a mysterious methodology since I&#8217;ve only seen their marketing copy rather than an arxiv paper, gets 90% on miniF2F 4 <em>when prompted with natural language proofs</em>. It doesn&#8217;t look to me like the deepseek or LEGO papers do that? I could be wrong. It&#8217;s impressive just to autoformalize natural language proofs, I guess I&#8217;m still wrapping my head around how much harder it is (for an LLM) to implement coming up with the proof as well.&nbsp;</p><p><a href="https://jobs.lever.co/harmonic">Jobs</a>: research engineer and software engineer</p><h1>Atlas ships their big google doc alluded to in the last newsletter</h1><p><a href="https://github.com/atlas-computing-org/awesome-AIxFV/blob/main/toolchain-for-ai-assisted-code-specification-synthesis-verification.pdf">Worth a read!</a> The GSAI stack is large and varied, and this maps out the different sub-sub-disciplines. From the executive summary:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!m7Qr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!m7Qr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 424w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 848w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 1272w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!m7Qr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png" width="1381" height="553" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f172a678-a574-442a-8514-6b4ed43f9469_1381x553.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:553,&quot;width&quot;:1381,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:383053,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!m7Qr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 424w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 848w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 1272w, https://substackcdn.com/image/fetch/$s_!m7Qr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff172a678-a574-442a-8514-6b4ed43f9469_1381x553.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You could start whole organizations for every row in this table, and I wouldn&#8217;t be a part of any org that targets more than a few at once for fear of being unfocused. See the doc for more <em>navigation</em> (see what I did there? Navigating like with an <em>atlas</em>, perhaps? Get it?) of the field&#8217;s opportunities.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><h1>Efficient shield synthesis via state-space transformation</h1><p>Shielding is an area of reactive systems and reinforcement learning that marks states as unsafe and synthesizes a kind of guarding layer between the agent and the environment that prevents unsafe actions from being executed in the environment. So in the rejection sampling flavored version, it literally intercepts the unsafe action and tells the agent &#8220;we&#8217;re not running that, try another action&#8221;. One of the limitations in this literature is computational cost, shields are, like environments, state machines plus some frills, and there may simply be too many states. This is the limitation that this paper focuses on.&nbsp;</p><blockquote><p><em><a href="https://arxiv.org/abs/2407.19911">We consider the problem of synthesizing safety strategies for control systems, also known as shields. Since the state space is infinite, shields are typically computed over a finite-state abstraction, with the most common abstraction being a rectangular grid. However, for many systems, such a grid does not align well with the safety property or the system dynamics. That is why a coarse grid is rarely sufficient, but a fine grid is typically computationally infeasible to obtain. In this paper, we show that appropriate state-space transformations can still allow to use a coarse grid at almost no computational overhead. We demonstrate in three case studies that our transformation-based synthesis outperforms a standard synthesis by several orders of magnitude. In the first two case studies, we use domain knowledge to select a suitable transformation. In the third case study, we instead report on results in engineering a transformation without domain knowledge.</a></em></p></blockquote><p>Besides cost, demanding a lot of domain knowledge is another limitation of shields, so this is an especially welcome development.</p><h1>Funding opportunities</h1><p><a href="https://www.aria.org.uk/programme-safeguarded-ai/">ARIA jumped right to technical area three (TA3)</a>, prototyping the gatekeeper. Deadline October 2nd. Seems geared toward cyber-physical systems folks. In the document:</p><blockquote><p><em>Note that verified software systems is an area which is highly suitable for a simplified gatekeeper workflow, in which the world-model is implicit in the specification logic. However, in the context of ARIA&#8217;s mission to &#8220;change the perception of what&#8217;s possible or valuable,&#8221; we consider that this application pathway is already perceived to be possible and valuable by the AI community. As such, this programme focuses on building capabilities to construct guaranteed-safe AI systems in cyber-physical domains. That being said, if you are an organisation which specialises in verified software, we would love to hear from you outside of this solicitation about the cyber-physical challenges that are just at the edge of the possible for your current techniques.</em></p></blockquote><p>This is really cool stuff, I hope they find brave and adventurous teams. I had thought gatekeeper prototypes would be in minecraft or mujoco (and asked a funder if they&#8217;d support me in doing that), so it&#8217;s wild to see them going for actual cyberphysical systems so quickly.&nbsp;</p><h1>Paper club</h1><p><a href="https://calendar.app.google/f7vR3f76z8g5UdPq5">Add to your calendar</a>. On September 19th we will read <a href="https://arxiv.org/abs/1810.08303">a paper about assume-guarantee contracts with learned components</a>. I&#8217;m liable to have made a summary slide deck to kick us off, but if I don&#8217;t, we&#8217;ll quietly read together for the first 20-30 minutes then discuss. The google meet room in the gcal event by default.&nbsp;&nbsp;</p><h1>Andrew Dickson&#8217;s excellent post</h1><p>See <a href="https://www.lesswrong.com/posts/B2bg677TaS4cmDPzL/limitations-on-formal-verification-for-ai-safety">Limitations on Formal Verification for AI Safety</a> over on LessWrong. I have a lot of agreements, and my disagreements are more a matter of what deserves emphasis than the fundamentals. Overall, I think the Tegmark/Omohundro paper failed to convey a <a href="https://en.wikipedia.org/wiki/Swiss_cheese_model">swisscheesey</a> worldview, and sounded too much like &#8220;why not just capture alignment properties in &#8216;specs&#8217; and prove the software &#8216;correct&#8217;?&#8221; (i.e. the vibe I was responding to in my <a href="https://www.lesswrong.com/posts/Egw5M6cnwgpRBXAQ4/speedrunning-4-mistakes-you-make-when-your-alignment">very pithy post</a>). However, I think my main reason I&#8217;m not using Dickson&#8217;s post as a reason to just pivot all my worldview and resulting research is captured in one of Steve&#8217;s comments:</p><blockquote><p>I'm focused on making sure our <em>infrastructure</em> is safe against AI attacks.</p></blockquote><p>Like, a very strong version I almost endorse is &#8220;GSAI isn&#8217;t about AI at all, it&#8217;s about systems coded by extremely powerful developers (which happen to be AIs)&#8221;, and ensuring safety, security, and reliability capabilities scale at similar speeds with other kinds of capabilities.</p><p>It looks like one can satisfy Dickson just by assuring him that GSAI is a part of a swiss cheese stack, and that no one is messianically promoting One Weird Trick To Solve Alignment. Of course, I do hope that no one is messianically promoting One Weird Trick&#8230;</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>One problem off the top of my head regarding the InterFramework section: Coq and Lean seems the most conceptually straightforward since they have the same underlying calculus, but even there just a little impredicativity or coinduction could lead to extreme headaches. Now you can have a model at some point in the future that steamrolls over these headaches, but then you have a social problem of the broader Lean community not wanting to upstream those changes&#8211; various forks diverging fundamentally seems problematic to me, would lead to a lot of duplicated work and missed opportunities for collaboration. I plan to prompt Opus 3.5 with &#8220;replicate <a href="https://gitlab.inria.fr/flocq/flocq">flocq</a> in lean4&#8221; as soon as I get access to the model, but how much more prompting effort will it be to ensure compliance with preexisting abstractions and design patterns, so that it can not only serve my purposes but be accepted by the community? At least there&#8217;s no coinduction in flocq, though some of the proofs may rely on set impredicativity for all I know (I haven&#8217;t looked at it in a while).</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[July 2024 Welcome Edition]]></title><description><![CDATA[A new newsletter about guaranteed safe AI]]></description><link>https://newsletter.for-all.dev/p/july-2024-welcome-edition</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/july-2024-welcome-edition</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Thu, 25 Jul 2024 20:35:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e6f92c2-2ffc-4b0b-ad92-596137aa1263_576x576.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>Welcome to the newsletter!&nbsp;</h1><p>Here we&#8217;ll be doing a monthly news roundup inclusive of my biased and wrong takes, plus maybe a remark about a non-current paper I happened to have read this month.&nbsp;</p><p>In this welcome edition, I won&#8217;t attempt to summarize all the news I missed before the newsletter existed, just the most recent notes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Progress in Guaranteed Safe AI to your inbox</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>There are no benefits to subscribing, the free tier has it all. But I enabled paid subscriptions just in case people want to signal support, to check how it affects my motivation to write these, and to see what happens.&nbsp;&nbsp;</p><p>Let me know changes I can make to maximize the impact of the newsletter by commenting here or emailing quinndougherty92 &lt;at&gt; gmail.</p><p>Thanks to Ben Goldhaber for the baseball metaphor idea and reviewing this post, and thanks to Steve Omohundro for suggesting the newsletter&#8217;s name, but all demerits go to me (the points go to me too btw).</p><h2>Where are we? What is guaranteed safe AI?&nbsp;</h2><p>Guaranteed-safe AI is a collection of research agendas to make the technological transition go well. There&#8217;s a video explainer <a href="https://www.youtube.com/watch?v=mwUfAvzjd8o">here</a>, but the main resource is the preprint position paper <a href="https://arxiv.org/abs/2405.06624">Towards Guaranteed Safe AI: A Framework for Ensuring Robust and Reliable AI Systems</a>.&nbsp;</p><p>Personally, I had <a href="https://forum.effectivealtruism.org/posts/4rMxiyPTPdzaFMyGm/high-impact-careers-in-formal-verification-artificial">a few years of telling people that formal verification is not the droid we&#8217;re looking for</a> before the latest push for GSAI, so you might be wondering if my worldview changed. In my earlier writing, I was mostly responding to a very strong &#8220;just capture alignment properties in &#8216;specs&#8217; and prove the implementation &#8216;correct&#8217;&#8221;, however I&#8217;m attracted to GSAI proposals because they have a <a href="https://en.wikipedia.org/wiki/Swiss_cheese_model">swiss cheese</a> matter-of-degree mindset that I don&#8217;t usually see when people get optimistic about the stronger version. Separately, proof assistants have played a big role in my career, so I&#8217;m biased toward theories of change that might leverage my experience.&nbsp;</p><h1>Atlas Computing talk</h1><p><a href="https://youtu.be/RPiLCy31d1E?si=Rf1r7wT8Xx8z2PqG">The monthly seminar featured Evan Miyazono of Atlas Computing, which is available on youtube</a>. The GSAI position papers are fine, but they need to be injected with a <em>roadmap, blueprint</em>, or <em>next steps</em>. Ideally, such a roadmap would go all the way down to estimates of headcount needed for different tactics. Wouldn&#8217;t it be great if someone was doing this? Anyways, major thanks to Atlas for doing this! Expect soon a document from Atlas that&#8217;s been incubating as a google doc all month.&nbsp;</p><h1>Podcast episode&nbsp;</h1><p><a href="https://www.cognitiverevolution.ai/guaranteed-safe-ai-world-models-safety-specs-verifiers-with-nora-ammann-ben-goldhaber/">Nora Ammann and Ben Goldhaber appeared on the Cognitive Revolution podcast</a> to discuss GSAI. The junction of world models, specifications, and assured implementations is so lofty and hairy that good science communication isn&#8217;t easy, so we need to see more of this.&nbsp;</p><h1>QAISI launch with job description</h1><p>The <a href="https://qaisi.ai/">Quantitative Safety AI Initiative</a> has a website up with a job posting for an AIS Research Engineer</p><p>The <a href="https://docs.google.com/forms/d/1irecYvgyrSR-5r8hN8qAXumheEuz02TtRgFp-QxYheY/edit">job</a> says it will mostly be focusing on <a href="https://arxiv.org/abs/2309.01933">the Tegmark &amp; Omohundro preprint</a>, even though two additional authors of <a href="https://arxiv.org/abs/2405.06624">the position paper of mammoth author list</a> are research leads for the orgs. <a href="https://www2.eecs.berkeley.edu/Faculty/Homepages/song.html">Dawn Song</a>, the one QAISI research lead <em>not</em> on that position paper, does a lot of security and cryptography work, but is no stranger to AI safety (she co-advised <a href="https://www.aisafetybook.com/">Dan Hendrycks</a>, off the top of my head). Excited to see what the new org gets up to!&nbsp;</p><h1>Guarantees-based mechanistic interpretability team preprint</h1><p>This paper isn&#8217;t 100% relevant, since GSAI tends more toward formally verifying artifacts of models than formally verifying models. However, it represents a natural thought you&#8217;d have when you try smashing formal verification into machine learning so I&#8217;m including it. This is not a commitment to make the newsletter a big tent for <em>any</em> FV/ML crossover episode overall, but it can be sometimes. <a href="https://arxiv.org/abs/2406.11779">Gross et. al.</a> lay the groundwork for formal proofs of model performance (lower bounds of accuracy). They accomplish this with mechinterp tricks like ablations, but since we&#8217;re dealing in proof we need<em> pessimal ablations</em> or ablating with worse cases rather than mean or random ablations.&nbsp;</p><p>I say &#8220;groundwork&#8221; because while the proofs in the current paper are &#8220;formal&#8221; by <em>many </em>standards, they are not yet formal in the sense of <em>machine checked</em>. But one step at a time, the current work seems like a big leap forward to me, and I&#8217;m skeptical of the marginal value add of putting those proofs in Coq.&nbsp;</p><p>See this <a href="https://www.alignmentforum.org/posts/bRsKimQcPTX3tNNJZ/compact-proofs-of-model-performance-via-mechanistic">splainer / companion piece by the authors</a> plus <a href="https://www.alignmentforum.org/posts/SyeQjjBoEC48MvnQC/formal-verification-heuristic-explanations-and-surprise">this supplemental by Jacob Hilton of ARC</a> (which funded the work).</p><h1>Progress in automating mathematics</h1><p>This is a <a href="https://arxiv.org/abs/2407.10040">very</a> <a href="https://arxiv.org/abs/2407.03203">active</a> <a href="https://arxiv.org/abs/2403.18120">area</a> with either tangential or moonshot safety stories, so I&#8217;ll never be exhaustive in covering the latest happenings, but how about a tasting.&nbsp;</p><h2>AlphaProof</h2><p>Deepmind: <a href="https://deepmind.google/discover/blog/ai-solves-imo-problems-at-silver-medal-level/">AI achieves silver-medal standard solving International Mathematical Olympiad problems</a>. Coupling a language model specializing in Lean with the AlphaZero RL algorithm, AlphaProof is competitive with IMO silver medalists, and <em>very</em> close to gold. The blog doesn&#8217;t link to a preprint, omits details about how the IMO data was Leanified, and omits the proportion of synthetic and self-play data vs other data. So it remains to be seen how significant this will be&#8211; I don&#8217;t have a strong inside view about <a href="https://arxiv.org/abs/2407.12220">benchmark hacking, but I know enough</a> about the problem area to be skeptical of everything.&nbsp;</p><p><a href="https://en.wikipedia.org/wiki/Curry%E2%80%93Howard_correspondence">Remember that proofs are programs</a>. We should expect <em>some</em> framing of the automated mathematics problem to do the conceptual heavy lifting for the automated software engineering problem. Reasons you&#8217;d hesitate to <a href="https://manifold.markets/CarsonGale/will-the-us-software-engineering-pr">buy up all the yes right now</a> is because IMO in lean is not that framing. <a href="https://proofassistants.stackexchange.com/questions/1115/how-usable-is-lean-for-constructive-mathematics">Lean (especially Mathlib) is, contra your assumptions when you walk into a dependent type theory, optimized for classical mathematics</a>. Without an analysis of the constructive capacities of AlphaProof&#8217;s capabilities, you shouldn&#8217;t assume that proofs are programs, machine learning may not walk down the halls of Curry-Howard just yet.&nbsp;</p><h1>GSAI paper club?&nbsp;</h1><p>I&#8217;m considering a monthly meetup to eat through papers. Comment here or email quinndougherty92 &lt;at&gt; gmail to express interest.&nbsp;</p><p>Shortlist of papers we could cover (not all of them timely)</p><ul><li><p><a href="https://arxiv.org/abs/2407.03203">TheoremLlama</a></p></li><li><p><a href="https://arxiv.org/abs/2306.04031">Certified Deductive Reasoning with LLMs</a></p></li><li><p><a href="https://arxiv.org/abs/2303.03226">RL with shields, but on ProbLog</a></p></li><li><p><a href="https://arxiv.org/abs/1810.08303">Assume-guarantee contracts with learned components</a></p></li></ul><p>Feel free to suggest others. Much of me would like to do a centralized summarizer (often me) format, but otherwise we could do a longer meetup where we read together on the call before discussing.&nbsp;</p><h1>Funding opportunities (jk)&nbsp;</h1><p>ARIA is currently closed, and SFF (which has some GSAI folks as panelists this year) deadline was a few weeks ago.&nbsp;</p><p>Neel Nanda, <a href="https://manifund.org/NeelNanda">leading a manifund budget</a>, informed me over discord that he&#8217;s pessimistic about GSAI and is unlikely to make grants in the area. I don&#8217;t know what other Manifund regrantors think, but there&#8217;s still a million or so in play for AI on manifund so could be worth a shot.&nbsp;</p><h1>Coming up</h1><p>You can <a href="https://lu.ma/xpf046sa">sign up for August&#8217;s GSAI seminar here</a>, from the authors of <a href="https://www.lesswrong.com/posts/y9tnz27oLmtLxcrEF/constructability-plainly-coded-agis-may-be-feasible-in-the">this post</a>. We&#8217;ll have to see what progress has been made in any prototyping that&#8217;s been done&#8211; the post I found to be pretty pie-in-the-sky.&nbsp;</p><h1>See you next month</h1><p>Like and subscribe, etc. Also, I am not of the twitterati so I don&#8217;t know how to hang out in specific places to make this newsletter better, but it&#8217;d be cool to cite other peoples&#8217; quick takes from time to time. To do this, I need your twitter account recommendations, so shoot those over.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Progress in Guaranteed Safe AI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[This is Can We Secure AI With Formal Methods?.]]></description><link>https://newsletter.for-all.dev/p/coming-soon</link><guid isPermaLink="false">https://newsletter.for-all.dev/p/coming-soon</guid><dc:creator><![CDATA[Quinn Dougherty]]></dc:creator><pubDate>Wed, 17 Jul 2024 21:28:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ykg_!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F083283b3-e660-4c7f-81e3-c40b1d1ebafd_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is Can We Secure AI With Formal Methods?.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.for-all.dev/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.for-all.dev/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>