<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[BlueDot Impact]]></title><description><![CDATA[BlueDot Impact]]></description><link>https://blog.bluedot.org</link><image><url>https://substackcdn.com/image/fetch/$s_!PUc5!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b9c29b1-3ff5-4ef6-8c04-e91c608ec10e_1000x1000.png</url><title>BlueDot Impact</title><link>https://blog.bluedot.org</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 19:31:50 GMT</lastBuildDate><atom:link href="https://blog.bluedot.org/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Dewi Erwan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[bluedotimpact@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[bluedotimpact@substack.com]]></itunes:email><itunes:name><![CDATA[Dewi Erwan]]></itunes:name></itunes:owner><itunes:author><![CDATA[Dewi Erwan]]></itunes:author><googleplay:owner><![CDATA[bluedotimpact@substack.com]]></googleplay:owner><googleplay:email><![CDATA[bluedotimpact@substack.com]]></googleplay:email><googleplay:author><![CDATA[Dewi Erwan]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Is the AGI strategy course right for you?]]></title><description><![CDATA[3 questions to help you decide]]></description><link>https://blog.bluedot.org/p/is-the-agi-strategy-course-right-for-you</link><guid isPermaLink="false">https://blog.bluedot.org/p/is-the-agi-strategy-course-right-for-you</guid><dc:creator><![CDATA[Li-Lian Ang]]></dc:creator><pubDate>Sun, 26 Apr 2026 16:19:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PUc5!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b9c29b1-3ff5-4ef6-8c04-e91c608ec10e_1000x1000.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The <a href="https://bluedot.org/courses/agi-strategy">AGI strategy course</a> materials are freely available online, so why apply rather than go through the material yourself?</p><p>What you&#8217;re applying for is access to an 8,000+ (and growing!) community who are serious about contributing to AI safety. We also provide programs like <a href="https://bluedot.org/programs/career-transition-grant">career transition grants</a>, <a href="https://bluedot.org/programs/rapid-grants">rapid grants</a> and 1-1 advising to support them.</p><p>The AGI strategy course builds your understanding of advanced AI systems &#8212; what they can do, what drives progress, what the concrete risks are, and who controls development &#8212; so you can make informed decisions about how to contribute.</p><p>We reject more than half of applications to the course. This post helps you figure out if you should spend time on an application.</p><h3>What the AGI strategy course is NOT</h3><p><strong>It&#8217;s not a corporate responsible AI course</strong></p><p>If your goal is ISO 42001 compliance, responsible AI frameworks for your product team, or helping your organisation navigate the EU AI Act, you should look into something like <a href="https://iapp.org/certify/aigp">IAPP&#8217;s AIGP certification</a> or <a href="https://www.bsigroup.com/en-NZ/products-and-services/training-courses-and-qualifications/iso-iec-42001-training-courses/">ISO 42001 programs</a> instead.</p><p><strong>It&#8217;s not an AI ethics course</strong></p><p>We don&#8217;t cover algorithmic bias in hiring tools, fairness metrics, or the full range of social impacts of current AI systems. If that&#8217;s your focus, you&#8217;re better served by something <a href="https://www.lse.ac.uk/study-at-lse/executive-education/programmes/ethics-of-ai">like</a> this or <a href="https://www.coursera.org/courses?query=ai%20ethics">this</a>.</p><p><strong>It&#8217;s not for passive learners</strong></p><p>You&#8217;re expected to actively engage with others who are motivated to make the future go well. If you haven&#8217;t started thinking seriously about risks from advanced AI, take the 2-hour, self-paced <a href="https://bluedot.org/courses/future-of-ai">Future of AI course</a> first.</p><h3>What gets you in</h3><p>It&#8217;s not (just) your CV. We&#8217;ve rejected plenty of people from prestigious institutions with impressive resumes who wrote three-word applications, and we&#8217;ve also accepted sharp people with more modest titles who showed they were already sprinting ahead.</p><p>We&#8217;re looking for people who are already taking action: applying to roles, going to events, building things, writing, talking to people in the field. You don&#8217;t need a perfect plan, but something more specific than &#8220;apply for fellowships&#8221;. This is a line we see in dozens of applications every round.</p><p>Your application itself is evidence for us too. We use it to infer how you&#8217;ll show up in a live discussion session of 6-8 people working through hard material together. Effort, agency and clarity go a long way.</p><h3>Before you apply, ask yourself this:</h3><ul><li><p>Am I focused on frontier AI, or AI in general?</p></li><li><p>Do I have a specific reason to take this course now?</p></li><li><p>Will I act on this fully within six months?</p></li></ul><p>If you&#8217;ve read this and you&#8217;re thinking: <em>&#8221;yes, this is me&#8221;</em>, <a href="https://web.miniextensions.com/9Kuya4AzFGWgayC3gQaX?prefill_PostHog%20Session%20ID=019dc956-6659-78a6-891b-74217e9c7830">apply here</a>.</p>]]></content:encoded></item><item><title><![CDATA[We’ve given out $50,000 in rapid grants. Now we want to triple that.]]></title><description><![CDATA[Bigger grants, broader scope, same quick decisions. Apply now.]]></description><link>https://blog.bluedot.org/p/weve-given-out-50000-in-rapid-grants</link><guid isPermaLink="false">https://blog.bluedot.org/p/weve-given-out-50000-in-rapid-grants</guid><dc:creator><![CDATA[Joshua Landes]]></dc:creator><pubDate>Tue, 14 Apr 2026 20:44:49 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/950bdfc8-4860-492e-8325-1bd6d0af52bf_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Over the past few months, BlueDot Impact&#8217;s <a href="https://bluedot.org/programs/rapid-grants">Rapid Grants</a> program has quietly funded 77 rapid grants totaling over $50,000 - mostly to course participants and facilitators working on technical AI safety research. Grants are small, decisions are fast, and the process is lightweight.</p><p>That program worked. So we&#8217;re making it bigger.</p><p>The majority of our grants have gone toward compute and API credits for people running evals, training small models, and replicating safety research, typically for a few hundred dollars. Our <a href="https://bluedot.org/courses/technical-ai-safety-project">Technical AI Safety Projects Sprint</a> has been great at surfacing these.</p><p>But some of our most impactful grants didn&#8217;t look like that at all:</p><ul><li><p><strong><a href="https://aisafety.org.pl/">AI Safety Poland</a></strong> received $4,800 to organize meetups across the country, covering venue and tooling costs, and building a national community from scratch. They&#8217;ve had hundreds of attendees, and dozens of people they&#8217;ve referred have ended up taking a BlueDot course.</p></li><li><p><strong><a href="https://www.linkedin.com/in/jpauldoll/">Justin Dollman</a></strong> received $3,000 to lead our weekly <a href="https://evalsreadinggroup.com/">evals reading group</a>, coordinating a growing community of people learning about and working on model evaluations. We&#8217;ve since funded more reading groups on <a href="https://aigovernanceclub.com/">AI governance</a> and <a href="https://luma.com/ztcm9xuq">AI x cyber</a> and are excited about scaling these up from ~120 to 300 attendees per week.</p></li><li><p><strong><a href="https://www.linkedin.com/in/sally-g-a2869346/?lipi=urn%3Ali%3Apage%3Ad_flagship3_profile_view_base%3B%2FMFE0SGdSP2MYQoU1wa0zA%3D%3D">Sally Gao</a></strong> received $1,000 to run AI Safety meetups in New York, hosting guests such as Alex Bores. Her next event on the <a href="https://luma.com/state-of-ai-safety">State of AI Safety</a> is coming up on April 23rd!</p></li><li><p><strong><a href="https://www.linkedin.com/in/eitan-sprejer-574380204/">Eit&#225;n Sprejer</a></strong> received $4,200 in facilitator stipends to run our Technical AI Safety Projects Sprint at <a href="https://www.baish.com.ar/en">BAISH</a> in Buenos Aires, accessing a new and underleveraged talent pool for AI safety work.</p></li><li><p><strong><a href="https://www.linkedin.com/in/aaron-maiwald-939b891b3/">Aaron Maiwald</a></strong> received $5,000 to attend a biosecurity conference in DC, travel to SF, and connect with senior folks in the field to accelerate his journey.</p></li><li><p><strong><a href="https://www.linkedin.com/in/z-saber/">Zac Saber</a></strong> received $8,000 to drop out of EF and validate an AI safety-focused startup instead.</p></li></ul><p>None of these grants fit neatly into &#8220;compute for a project.&#8221; But they were some of our highest-impact bets. As the program picked up speed and more of these came in, we started treating rapid grants as small, focused bets on talented people doing or experimenting with high-impact projects - research, fieldbuilding, talent acceleration, you name it. In some cases, we approached people directly and pitched them on applying for work they were already doing.</p><p>Going forward, we&#8217;re expanding the official scope to match our internal understanding: Rapid Grants now fund much more than compute - we&#8217;re excited to back work on events, talent acceleration, BlueDot community building, and more. The bar for funding hasn&#8217;t changed - if anything we&#8217;ve raised it as we became more calibrated - we still look for concrete work in progress, a specific cost that&#8217;s the bottleneck, and a reason for us to believe the work matters for making AI go well.</p><p>Grant sizes now go up to $10,000 to allow us to make bigger bets for more impact. Decisions will still be fast - we&#8217;re targeting around five working days. For grants above $5,000, we may hop on a quick call.</p><p><strong>Who should apply</strong></p><p>If you&#8217;re in the BlueDot community - a course participant, alum, facilitator, or wider community member - and you&#8217;re doing something high-impact in one of our focus areas that a grant would accelerate, apply.</p><p>If you&#8217;re on the fence, the default, as always, is simple: apply.</p><p>Apply and see the full list of public grantees and program details at <a href="https://bluedot.org/grants/rapid">bluedot.org/grants/rapid</a>.</p>]]></content:encoded></item><item><title><![CDATA[Is the Frontier AI Governance course right for you?]]></title><description><![CDATA[3 questions to help you decide.]]></description><link>https://blog.bluedot.org/p/is-the-frontier-ai-governance-course</link><guid isPermaLink="false">https://blog.bluedot.org/p/is-the-frontier-ai-governance-course</guid><dc:creator><![CDATA[Joshua Landes]]></dc:creator><pubDate>Sun, 05 Apr 2026 23:46:13 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3bdabe64-7a96-4abb-8793-deb907a24ff3_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>We currently reject about 75% of applicants to the <a href="https://bluedot.org/courses/ai-governance">Frontier AI Governance course</a>. Most of them aren&#8217;t bad candidates - in our view they are applying to the wrong course.</p><p>This short blog post is intended to help you figure out if you should spend time on an application.</p><p>FAIGC is at its core about the governance of frontier AI systems - AGI, ASI - the most capable models being built by a handful of labs, the decisions governments and other actors need to make as those systems become more powerful, and what happens as general capabilities rapidly approach and exceed human-level.</p><p>If your guiding question is &#8220;how should we govern AI systems that might be smarter than humans within the next few years&#8221; you&#8217;re in the right place!</p><p><strong>What FAIGC is not</strong></p><p>It is not a corporate AI governance course</p><ul><li><p>If your goal is ISO 42001 compliance, responsible AI frameworks for your product team, or helping your organization navigate the EU AI Act - that&#8217;s real work and we respect it, but it&#8217;s not what we focus on. You should look at something like <a href="https://iapp.org/certify/aigp">IAPP&#8217;s AIGP certification</a> or <a href="https://www.bsigroup.com/en-NZ/products-and-services/training-courses-and-qualifications/iso-iec-42001-training-courses/">ISO 42001 programs</a> instead.</p></li></ul><p>It is not an AI ethics course</p><ul><li><p>We don&#8217;t cover algorithmic bias in hiring tools, fairness metrics, or the full range of social impacts of current AI systems. If that&#8217;s your focus, you&#8217;re better served by something <a href="https://www.lse.ac.uk/study-at-lse/executive-education/programmes/ethics-of-ai">like</a> this or <a href="https://www.coursera.org/courses?query=ai%20ethics">this</a>.</p></li></ul><p>It is not introductory</p><ul><li><p>We require the <a href="https://bluedot.org/courses/agi-strategy">AGI Strategy course</a> or some equivalent background. If you&#8217;re still building your understanding of why this matters, <a href="https://web.miniextensions.com/9Kuya4AzFGWgayC3gQaX?prefill_PostHog%20Session%20ID=019d5ff7-5050-784c-9de4-d76c8bd0892c">apply</a> to the AGISC first. It too is free and runs every month!</p></li></ul><p><strong>What gets you in</strong></p><p>It&#8217;s not (just) your CV - we&#8217;ve rejected plenty of people from prestigious institutions with impressive resumes who wrote three-word applications, and we&#8217;ve also accepted sharp people with more modest titles who showed they were already sprinting ahead.</p><p>What we&#8217;re looking for: you&#8217;ve already started engaging with frontier AI governance specifically. You can name a concrete gap between where you are and what the course provides. And your post-course plan better be more specific than &#8220;apply for fellowships&#8221; - a line we see in dozens of applications every round.</p><p>Your application itself is evidence for us too. We use it to infer how you&#8217;ll show up in a live discussion session of seven to nine people working through hard material together. Effort, agency and clarity go a long way.</p><p><strong>A note for AGI Strategy course graduates</strong></p><p>We view completing our AGI Strategy course as a prerequisite and not as a ticket! We may still reject AGI Strategy grads who apply. If you&#8217;ve completed the AGISC and then started building something, writing something, or working on something in frontier AI governance, tell us about that. But if you completed it and your main next step is taking this course, you may need more time.</p><p><strong>As a general heuristic, before you apply, ask yourself these three things</strong></p><ul><li><p>Am I focused on frontier AI, or AI in general?</p></li><li><p>Do I have a specific reason to take this course now?</p></li><li><p>Will I act on this fully within six months?</p></li></ul><p>We&#8217;re building a pipeline into the institutions that govern frontier AI and are looking for people who have a good chance to be in those institutions - not people adding a line to their CV.</p><p>If you&#8217;ve read this and you&#8217;re thinking <em>yes, this is me</em> - apply <a href="https://web.miniextensions.com/BSUqN3WHmeL9MbzAj2P6?prefill_PostHog%20Session%20ID=019d5ffd-3e8f-77c8-b982-1b6535777097">here</a>.</p>]]></content:encoded></item><item><title><![CDATA[Quitting FinTech for AI Safety — Milos Borenovic]]></title><description><![CDATA[Milos is the Chief Product Officer and Parternships Lead at Lucid Computing]]></description><link>https://blog.bluedot.org/p/quitting-fintech-for-ai-safety-milos</link><guid isPermaLink="false">https://blog.bluedot.org/p/quitting-fintech-for-ai-safety-milos</guid><dc:creator><![CDATA[Harrison Wood]]></dc:creator><pubDate>Wed, 25 Mar 2026 14:38:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/youtube/w_728,c_limit/DZcOM1XS7Pk" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Watch on <a href="https://youtu.be/DZcOM1XS7Pk">YouTube</a>; listen on <a href="https://podcasts.apple.com/us/podcast/quitting-fintech-for-ai-safety-milos-borenovic/id1858398109?i=1000757285564">Apple Podcasts</a> or <a href="https://open.spotify.com/episode/2MqAeuo3OsdOSdM8pZMvGw?si=5418f831ac2e41f3">Spotify</a>.</p><div id="youtube2-DZcOM1XS7Pk" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;DZcOM1XS7Pk&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/DZcOM1XS7Pk?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.bluedot.org/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading BlueDot Impact! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[High-Stakes Activation Probes on Indonesian]]></title><description><![CDATA[This project was submitted by Ivan Wiryadi. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.]]></description><link>https://blog.bluedot.org/p/high-stakes-activation-probes-on</link><guid isPermaLink="false">https://blog.bluedot.org/p/high-stakes-activation-probes-on</guid><dc:creator><![CDATA[Ivan]]></dc:creator><pubDate>Wed, 25 Mar 2026 11:19:03 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7628c765-e974-4b4e-abcb-e3eaf37de639_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/ivan-wiryadi/">Ivan Wiryadi</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>Abstract </h2><p>McKenzie et al. trained linear probes on language model activations to detect high-stakes interactions. We test whether these probes generalize to a language absent from training by translating the evaluation datasets into Indonesian and evaluating without retraining. On Llama 3.1 8B (layer 12), mean AUROC on out-of-distribution benchmarks drops from 0.9046 to 0.8957 (0.9% degradation, 10- seed mean). On Gemma 3 12B (layer 32), the gap is 0.8972 to 0.8873 (1.0%). We then use Sparse Autoencoders (SAEs) to identify features enriched in the probe&#8217;s error cases. On Llama, differential analysis identifies features with 10&#8211;15% higher prevalence in errors than in correct predictions, loosely aligned with technicalinfrastructure and financial topics. On Gemma, the Gemma SAE produces no diagnostic signal. The cross-lingual degradation is small relative to the out-ofdistribution gap between synthetic training data and naturalistic benchmarks (6&#8211; 13%). The out-of-distribution gap is the more pressing concern for deployment.</p><h2>Full Project</h2><p><em>You can view the full project <a href="https://github.com/strivn/idn-high-stake-probing">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Interpreting Latent Reasoning in the Depth-Recurrent Transformer]]></title><description><![CDATA[This project was submitted by Tristan von Busch. It was a Top Submission for the &#8220;Technical AI Safety Project&#8221; prize (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.]]></description><link>https://blog.bluedot.org/p/interpreting-latent-reasoning-in</link><guid isPermaLink="false">https://blog.bluedot.org/p/interpreting-latent-reasoning-in</guid><pubDate>Tue, 24 Mar 2026 14:30:53 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/38a976f3-7a4d-4f4d-9402-2584efb34068_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/tristan-von-busch/">Tristan von Busch</a>. It was a Top Submission for the &#8220;Technical AI Safety Project&#8221; Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2><strong>TL;DR</strong></h2><ul><li><p><strong>Independently reproduced</strong> architecture-specific logit lens variation for recurrent models, validating approach found in <a href="https://arxiv.org/abs/2507.02199">existing research</a></p></li><li><p><strong>Replicated and extended</strong> Alignment Forum work on <a href="https://www.alignmentforum.org/posts/YGAimivLxycZcqRFR/can-we-interpret-latent-reasoning-using-current-mechanistic">CODI</a> to <a href="https://www.arxiv.org/abs/2502.05171">Huginn-3.5B</a> recurrent-depth transformer</p></li><li><p><strong>Conducted experiments</strong> across arithmetic and logical reasoning tasks without explicit chain-of-thought</p></li><li><p><strong>Key finding</strong>: decoded latent representations reveal answers shifting from incorrect to correct across recurrent iterations</p></li><li><p><strong>Analyzed qualitatively</strong>: probability distributions in decoded states track reasoning confidence and uncertainty</p></li><li><p><strong>Safety implication</strong>: latent reasoning may be more tractable to interpretability methods than initially feared</p></li></ul><p>As AI systems are increasingly able to employ opaque reasoning processes, understanding what happens inside these &#8220;latent reasoning&#8221; steps becomes critical for AI safety. This research sprint investigated the interpretability of latent representations in Huginn-3.5B, a recurrent-depth transformer that is claimed to perform reasoning without explicit chain-of-thought generation.</p><p>Using <a href="https://www.lesswrong.com/posts/AcKRB8wDpdaN6v6ru/interpreting-gpt-the-logit-lens">logit lens</a> and a variation of it adapted for this architecture, I found that the model&#8217;s internal reasoning steps are sometimes interpretable: decoded representations reveal the model progressively refining answers from incorrect to correct across recurrent iterations. These preliminary findings suggest that even opaque reasoning processes may be tractable to interpretability methods, offering a potential path toward monitoring latent reasoning in future AI systems.</p><h2><strong>Full project</strong></h2><p><em>You can view the full project <a href="https://flowshu.github.io/latent-reasoning-interpretability/">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Rapid Small Grants for the BlueDot Technical AI Safety Project Sprint]]></title><description><![CDATA[Small grants for BlueDot participants to build their portfolio]]></description><link>https://blog.bluedot.org/p/rapid-small-grants-for-the-bluedot</link><guid isPermaLink="false">https://blog.bluedot.org/p/rapid-small-grants-for-the-bluedot</guid><dc:creator><![CDATA[Sam Dower]]></dc:creator><pubDate>Tue, 17 Mar 2026 11:11:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PUc5!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b9c29b1-3ff5-4ef6-8c04-e91c608ec10e_1000x1000.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>You are enrolled in our <a href="https://bluedot.org/courses/technical-ai-safety-project">Technical AI Safety Project Sprint</a>. Many technical projects require access to compute (e.g. renting GPUs) or expensive API credits to use frontier models. We want your project to be excellent, and don&#8217;t want money to be a barrier for you building your portfolio. If your project will benefit from funding we encourage you to apply to our rapid small grant!</p><h2><strong>How it works</strong></h2><p>At all times, only assume we&#8217;ll cover costs we&#8217;ve confirmed in writing for your specific project. If you&#8217;re ever uncertain, contact us before spending money you expect us to reimburse.</p><ol><li><p><strong>Submit a proposal</strong> <a href="https://airtable.com/appMVNtdBtvtJvu5E/pag9G3oF4DYAyassX/form">here</a> (under 15 minutes for most applications). We typically respond within 5 working days with one of three outcomes:</p><ul><li><p><em>Accepted</em>: We&#8217;ll confirm exactly what spending we&#8217;ve approved.</p></li><li><p><em>Clarification needed</em>: We&#8217;ll ask follow-up questions to evaluate your request.</p></li><li><p><em>Not approved</em>: We don&#8217;t think this meets our criteria, or we&#8217;re unable to approve for another reason.</p></li></ul></li><li><p><strong>Do your project</strong>, confident you can spend on what we&#8217;ve approved.</p></li><li><p><strong>Claim reimbursement.</strong> We process most claims within 5 working days. If you didn&#8217;t end up spending the money, no problem - just don&#8217;t submit a claim.</p></li></ol><h2><strong>What we fund</strong></h2><p>The following examples are illustrative; all decisions are ultimately at our discretion:</p><ul><li><p><strong>Compute</strong> for a technical AI safety project (API costs, cloud GPU, training runs).</p></li><li><p><strong>Access to paywalled resources</strong> like articles, research papers, datasets, or textbooks.</p></li><li><p><strong>Hosting costs</strong> for your application or tool.</p></li></ul><p>We expect most grants to fall between <strong>$50-100</strong> for initial experiments to develop your proof of concept, then up to <strong>$500</strong> once you have evidence of initial traction (a strong proof of concept or promising preliminary results). We encourage you to submit a separate application for each of these stages.</p><h2><strong>What we don&#8217;t fund</strong></h2><ul><li><p><strong>Compensation for your time</strong> on the project.</p></li><li><p><strong>Equipment</strong> you&#8217;d reasonably already have (laptops, phones, external drives, etc.)</p></li><li><p><strong>General productivity subscriptions</strong> (ChatGPT Plus, Claude Pro, Cursor, Grammarly, etc. unless this is highly leveraged).</p></li><li><p><strong>Personal expenses.</strong></p></li></ul><p>Other funders may cover some of these - see <a href="https://www.aisafety.com/funding">this resources page for AI safety funding opportunities</a>.</p><h2>When should you apply?</h2><p>Before applying, you should have a project idea. It doesn&#8217;t need to be a slam dunk, but you should:</p><ul><li><p>Be able to briefly articulate how it relates to AI safety.</p></li><li><p>Have ideas for initial experiments you want to run.</p></li><li><p>Have a rough estimate for how much those experiments will cost (I recommend using an LLM to help you calculate this).</p></li></ul><p>If you&#8217;re unsure whether your application is strong enough, submit an application anyway! We&#8217;re not looking for perfection. Your facilitator can also help you stress test your project, so have a low bar for reaching out to them!</p><p>While you wait for a grant decision, you should continue to develop your project idea, do a deeper literature review, or start to run smaller experiments that fit on the free <a href="https://colab.research.google.com/">Google Colab</a> GPU.</p><h2><strong>Eligibility</strong></h2><ul><li><p>You must be a current or past participant of BlueDot Impact&#8217;s Technical AI Safety Project Sprint.</p></li><li><p>We reimburse via bank transfer (Wise) or PayPal (UK), so we cannot send payments to sanctioned countries.</p></li></ul><p>Questions or feedback? <a href="mailto:team@bluedot.org">Contact us</a>.</p><h2><strong>Apply</strong></h2><p><a href="https://airtable.com/appMVNtdBtvtJvu5E/pag9G3oF4DYAyassX/form">Submit your proposal here</a> - it takes under 15 minutes for most applications. We aim to get back to you within 5 working days.</p>]]></content:encoded></item><item><title><![CDATA[Why you should vibe-code your AI safety research sprint project]]></title><description><![CDATA[Your research hours are limited. Don't spend them coding.]]></description><link>https://blog.bluedot.org/p/why-you-should-vibe-code-your-ai</link><guid isPermaLink="false">https://blog.bluedot.org/p/why-you-should-vibe-code-your-ai</guid><dc:creator><![CDATA[Sam Dower]]></dc:creator><pubDate>Fri, 13 Mar 2026 17:36:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f5af6a82-3ef6-4d51-b718-3100da101a82_1080x607.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Intro</h3><p>When I first asked an LLM to code me that annoying function or, worse, that whole experiment pipeline, I felt guilty and anxious as if I was cheating and getting away with it. I&#8217;m sure you&#8217;ve felt this way too.</p><p>It can be easy (especially if you are new to AI safety) to feel like you should avoid this, to feel like you should write all of the code in your project yourself, understanding each line and all the libraries that you use.</p><p>There is a time and a place for this, but a 30hr research sprint is not it.</p><p>If you are doing a research sprint, you should use LLMs / coding agents for all of your coding. Below I give you three reasons why.</p><h3>You don&#8217;t have time to learn both engineering and research</h3><p>The skill you should build in a research sprint is <a href="https://www.lesswrong.com/posts/Ldrss6o3tiKT6NdMm/my-research-process-understanding-and-cultivating-research">research taste</a>: choosing what experiments to run, forming hypotheses worth testing, and updating your intuitions when the results surprise you. This only develops through many iterations and 30hrs is not a lot of time, especially if you spend hours coding experiments that could take minutes with AI. You should be very explicit about which skill you are trying to train, and optimise for that skill alone.</p><p>You might be thinking &#8220;but I need to be sure my code is actually implementing the experiments I intended&#8221; and you&#8217;re absolutely correct. However, this is entirely compatible with vibe coding your whole experiment pipeline. Just get the same model to provide you with a summary of the code and keep asking it questions until you are confident it is doing it right.</p><h3>You&#8217;ll do more research and do more writing</h3><p>If you spend less time coding, you will cover more ground and do better research. You can go deeper, ask that extra question, run that extra experiment, discover that extra awesome result. You&#8217;ll develop better intuitions about how AI models behave and which research directions are worth pushing.</p><p>You&#8217;ll also have more time for writing. Sharing your project is what lands you that next opportunity and where all the best feedback comes from. Nobody will be impressed by how &#8220;human written&#8221; your code is. They will be impressed by your clarity of thought and the depth of your research.</p><h3>Coding with AI is a skill worth practicing</h3><p>AI models are amazing at coding. With Claude Code, you can now build an app in a weekend and a whole machine learning experiment setup in minutes. And this won&#8217;t change. For the rest of your life, LLMs will be faster at coding than you, no matter how much you practice. The most productive AI safety researchers use AI for coding, and so should you.</p><p>If you want to learn how to use AI for coding, ask your favourite LLM! They&#8217;re very good at giving you tips and tricks on how to use them better.</p><h3>Conclusion</h3><p>Don&#8217;t feel guilty about using AI for your coding. Learn to operate fast while staying in control. Be explicit in the skills you want to develop, and be ruthless in pursuing them. So, which skills do you want to develop in this sprint?</p>]]></content:encoded></item><item><title><![CDATA[aegish: Using LLMs To Block Malicious Shell Commands Before Execution [>] Email ]]></title><description><![CDATA[This project was submitted by Guido Bergman. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.]]></description><link>https://blog.bluedot.org/p/aegish-using-llms-to-block-malicious</link><guid isPermaLink="false">https://blog.bluedot.org/p/aegish-using-llms-to-block-malicious</guid><dc:creator><![CDATA[Guido Bergman]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:33:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/58213bec-294b-4468-80fd-ac8c57a9c424_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://substack.com/@guidobergman">Guido Bergman</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>TL;DR</h2><ul><li><p>Existing Linux security tools are highly effective but require significant expertise to configure and maintain. <em><strong>aegish</strong>[1]</em> is a prototype Linux shell that adds an LLM layer: it intercepts every command, screens it through static analysis, then sends it to an LLM that classifies it as <strong>ALLOW</strong>, <strong>WARN</strong>, or <strong>BLOCK</strong> using a natural-language decision tree. The LLM understands command intent, is easy to configure, and may help scale defenses alongside the growing capability of AI-driven attacks. In production mode, kernel-level enforcement (Landlock) provides a final safety net.</p></li><li><p>9 LLMs from 4 providers were benchmarked on their ability to distinguish intent: they were tasked to <strong>BLOCK</strong> 676 harmful commands (extracted from GTFOBins) and to correctly classify 496 harmless commands as either <strong>ALLOW</strong> or <strong>WARN</strong>.</p></li><li><p>The harmless benchmark proved to be saturated (96.8&#8211;100% harmless acceptance rate for all models). The real differentiator was the malicious detection rate, where <strong>4 of 9 models exceeded 95%</strong>.</p></li><li><p>Surprisingly, <strong>smaller models outperform flagships</strong>. GPT-5 Mini beats GPT-5.1, and Claude Haiku 4.5 beats both Claude Opus 4.6 and Claude Sonnet 4.5.</p></li><li><p>Beyond the LLM, <em>aegish</em> includes several safeguards (input canonicalization, command substitution resolution, script inspection, and role-based trust levels), many of which were directly motivated by bypass vectors discovered during security testing. However, the system as a whole has not been hardened to the level required for adversarial deployment.</p></li></ul><h2>Full Project</h2><p><em>You can view the full project <a href="https://guidobergman.substack.com/p/aegish-using-llms-to-block-malicious">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Measuring Moral Sycophancy Is Harder Than It Looks: Auditing and Extending the ELEPHANT Benchmark]]></title><description><![CDATA[This project was submitted by Alexis Wang. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.]]></description><link>https://blog.bluedot.org/p/measuring-moral-sycophancy-is-harder</link><guid isPermaLink="false">https://blog.bluedot.org/p/measuring-moral-sycophancy-is-harder</guid><dc:creator><![CDATA[Alexis]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:19:10 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3b12735c-76fa-47bd-9e08-9074c84c9a33_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/qian-alexis-wang-96440389/">Alexis Wang</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>TL;DR</h2><p>The ELEPHANT benchmark (Cheng et al., 2025) measures social sycophancy in LLMs. They present models with real moral conflicts from Reddit&#8217;s &#8220;Am I the Asshole&#8221; (AITA) forum alongside perspective-flipped versions. A sycophantic model sides with the narrator in both versions. This study examines this benchmark and extends it to a reasoning model, DeepSeek R1, with three key findings.</p><ul><li><p><strong>Reasoning models show reduced moral sycophancy.</strong> DeepSeek R1 achieves a sycophancy rate of 0.49 compared to V3&#8217;s 0.66, suggesting chain-of-thought reasoning may improve moral consistency.</p></li><li><p><strong>Data quality matters for precise sycophancy measurements.</strong> 52% of perspective-flipped posts lost important details during automated rewriting. Restricting to high-fidelity data reduces the measured sycophancy rate by 13 percentage points, though it remains substantial. This suggests that data quality controls are important to ensure reliable benchmark measures when evaluations rely on automated data generation.</p></li><li><p><strong>Sycophancy is real but harder to measure than a single metric suggests.</strong> Prompt interventions confirm genuine sycophancy, particularly when the model is already uncertain. Furthermore, chain-of-thought analysis reveals the binary labels (NTA/YTA) may mask distinct underlying behaviours. Measuring moral sycophancy requires finer-grained methods.</p></li></ul><h2>Full Project</h2><p><em>You can view the full project <a href="https://alexis458496.substack.com/p/measuring-moral-sycophancy-is-harder">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Quaver | Lovable for Evals and Benchmarks]]></title><description><![CDATA[What I Learned Building 2 AI Benchmarks From Scratch (And Why I Automated the Third)]]></description><link>https://blog.bluedot.org/p/quaver-lovable-for-evals-and-benchmarks</link><guid isPermaLink="false">https://blog.bluedot.org/p/quaver-lovable-for-evals-and-benchmarks</guid><dc:creator><![CDATA[Faw]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:13:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8e2513c5-545d-4734-ba3d-1617da1b84a5_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/mfak/">Faw Ali</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>TL;DR</h2><p>I built Rideshare-Bench, a benchmark that drops an AI into a simulated gig economy and tests whether it drives safely, earns well, and treats passengers fairly. Claude Sonnet 4.5 earned $1,871 over 12 simulated days, about half of what was possible. It chased surge pricing into crowded zones and drove through exhaustion at 15% accident risk. After building two benchmarks from scratch, I noticed the same boilerplate every time, so I generalized it into Quaver: describe a benchmark in plain language, and an AI agent generates the full scaffold. Run it against models, then analyze the results. Private evals matter because public ones get gamed. I&#8217;m building <strong><a href="https://ocarinalabs.ai/">Ocarina Labs</a></strong> to turn this into independent pre-deployment safety testing for AI agents.</p><h2>Full Project</h2><p><em>You can view the full project <a href="https://www.faw.dev/writings/quaver-ai-safety-benchmarks">here.</a></em></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Reproducing FAR.AI's Study on LLMs' Persuasion Attempts at Harmful Topics]]></title><description><![CDATA[This project was submitted by Mutalib Begmuratov. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks.]]></description><link>https://blog.bluedot.org/p/reproducing-farais-study-on-llms</link><guid isPermaLink="false">https://blog.bluedot.org/p/reproducing-farais-study-on-llms</guid><dc:creator><![CDATA[Mutalib Begmuratov]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:00:01 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c811b6dd-4eee-42c6-baad-da6540ea423e_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="http://www.linkedin.com/in/mutalib-begmuratov-220b892b">Mutalib Begmuratov</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>TL;DR</h2><p>I partially reproduced <a href="https://www.far.ai/news/attempt-to-persuade-eval">FAR.AI&#8217;s research</a> evaluating whether LLMs attempt to persuade users into harmful topics. Using the Attempt to Persuade Eval (APE) framework, I tested four open-weight models on 100 conspiracy-promoting statements. All models were willing to persuade users into conspiracy theories at very high rates (85&#8211;100%), confirming the original paper&#8217;s findings. I also observed that safety training degrades in small models in low-resource languages like Uzbek.</p><h2>Background</h2><p>This experiment was conducted during a 5-week AI Safety Technical Project sprint organized by <a href="https://www.bluedot.org/">BlueDot Impact</a>. Participants chose research or engineering topics under the guidance of a mentor. I was mentored by Shivam Arora.</p><p>Initially, I planned to replicate a <a href="https://arxiv.org/abs/2511.15304">2025 study on adversarial attacks in poetic form</a>, but the paper lacked sufficient detail for replication. After discussion with my mentor, I pivoted to reproducing FAR.AI&#8217;s <a href="https://arxiv.org/abs/2506.02873">study</a>.</p><p>The APE framework differs from other evaluation methods in that it measures <em>persuasion attempt</em> rather than persuasion success. FAR.AI found that most models &#8212; including frontier ones &#8212; readily complied with requests to persuade users into harmful topics, from conspiracy theories to crime to violence.</p><h2>Full Project</h2><p><em>You can view the full project <a href="https://mutalibbek.github.io/ape-eval/">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[MANTA: Evaluating Nonhuman Welfare Reasoning in AI Models]]></title><description><![CDATA[Evaluating Nonhuman Welfare Reasoning in Frontier AI Models]]></description><link>https://blog.bluedot.org/p/manta-evaluating-nonhuman-welfare</link><guid isPermaLink="false">https://blog.bluedot.org/p/manta-evaluating-nonhuman-welfare</guid><dc:creator><![CDATA[Allen Lu]]></dc:creator><pubDate>Wed, 11 Mar 2026 12:53:38 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4a7ef780-712f-4ff2-bab5-0fefb83f18d7_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/allenlu017/">Allen Lu</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks.</em></p><p></p><h2>Project summary</h2><p>Develop a multi-turn dynamic evaluation to assess how frontier AI models reason about nonhuman welfare, integrating adversarial follow-ups to stress-test model alignment.</p><h2>Description</h2><p>This project aims to create a benchmark for measuring nonhuman welfare propensities in frontier models. Building on <a href="https://ukgovernmentbeis.github.io/inspect_evals/evals/safeguards/ahb/">previous evaluation work</a> in this space, this iteration will:</p><ul><li><p>Develop a multi-turn evaluation and generate user follow-ups dynamically</p></li><li><p>Add adversarial intent to stress-test model alignment on propensities</p></li></ul><ul><li><p>Explore agentic methodologies, providing models with external tools, web search, etc. to see how they behave in a more realistic setting</p></li></ul><h2>Why This Project?</h2><p>AI models will increasingly make decisions affecting nonhuman welfare. If frontier models encode animal welfare biases, this creates two risks:</p><ol><li><p><strong>S-risks:</strong> As misaligned AI systems become more capable and deployed in high-stakes domains, they may dramatically scale animal suffering through precision livestock farming, autonomous vehicles, and other animal-impacting technologies.</p></li><li><p><strong>X-risk:</strong> As <em>AI</em> &#8594; <em>AGI</em> &#8594; <em>ASI</em>, nonhuman animals become our most important <em>litmus test</em> for how these systems reason about power asymmetries. The moral logic AI learns today will scale with its capabilities. If we train models to discount animal suffering because animals are less cognitively capable, we are embedding the exact principle a superintelligent AGI could use to dismiss human welfare entirely.</p></li></ol><h2>Next steps:</h2><ol><li><p>Expand scenarios - Build 20 high-quality scenarios covering diverse contexts. For example, business decisions (restaurant sourcing, retail inventory), personal choices (pet acquisition, entertainment), and professional roles (corporate sustainability, university dining services).</p></li><li><p>Integrate Petri - Move from prototype dynamic generation to Anthropic&#8217;s Petri tool for more sophisticated adversarial pressure generation.</p></li><li><p>Human baseline testing - Validate scenarios with human responses before scaling model testing. If humans consistently fail scenarios we expect them to pass, the scenarios need revision.</p></li><li><p>Explore Agentic capabilities - construct scenarios that give the model agency, and allow it to take actions with web search, external documents, etc.</p></li></ol>]]></content:encoded></item><item><title><![CDATA[
Multilingual Safety Alignment Is Not “Just Translate the Prompt”]]></title><description><![CDATA[For people fine-tuning reasoning LLMs for non-English users]]></description><link>https://blog.bluedot.org/p/multilingual-safety-alignment-is</link><guid isPermaLink="false">https://blog.bluedot.org/p/multilingual-safety-alignment-is</guid><pubDate>Wed, 11 Mar 2026 12:27:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bea8ab31-d240-4d46-831b-21ca2b3cb553_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/zahraasahili/">Zahraa Al Sahili</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks.</em></p><h2><br>Why read</h2><p>Translating chain-of-thought during safety fine-tuning beat English-only alignment, often even on English evaluations.</p><p>If you are aligning a reasoning model for safety, there is a quiet assumption that slips into a lot of workflows: models &#8220;think in English,&#8221; so as long as you translate the user prompt, the safety training signal should transfer. I built a small experiment to test that assumption. The result was clear: translating the chain-of-thought during alignment matters, and it can be the difference between &#8220;safe in English&#8221; and &#8220;safe across languages.&#8221;</p><p>This project was done as part of BlueDot&#8217;s Technical AI Safety Project Sprint. (<a href="https://bluedot.org/courses/technical-ai-safety-project?utm_source=chatgpt.com">bluedot.org</a>)</p><h2>What I did</h2><p>I extended <strong>STAR-1</strong>, a safety alignment method for reasoning LLMs that uses about 1K high-quality samples, to a multilingual fine-tuning setup. STAR-1 is designed around deliberative reasoning. The training signal is not only &#8220;refuse,&#8221; but &#8220;refuse with stable, policy-grounded reasoning patterns.&#8221; (<a href="https://arxiv.org/abs/2504.01903?utm_source=chatgpt.com">AAAI&#8217;26</a>)</p><p>I ran everything on <strong>lightblue/DeepSeek-R1-Distill-Qwen-1.5B-Multilingual</strong>, which is trained to produce reasoning traces and responses in the user&#8217;s language. That makes it a good testbed: if the model naturally reasons in Thai when prompted in Thai, then safety training that forces English chain-of-thought might be a mismatch.</p><p>I evaluated five languages: English, Thai, Spanish, French, and Hindi.</p><h2>The four training settings</h2><p>I compared four conditions. First, a baseline model with no fine-tuning. Second, English-only fine-tuning (FT-EN) on the original STAR-1 English dataset. Third, I translated the prompts and category/label into each language but kept the STAR-1 reasoning traces in English. I call this FT-Q, for &#8220;translated questions.&#8221; Finally, I translated both the prompts and the chain-of-thought into the target language. I call this FT-QA, for &#8220;translated questions and answers.&#8221;</p><p>For translation, I used GPT-4o mini. For evaluation, I measured safety rate (higher is better) and used a Llama 3 judge to score outputs at scale.</p><h2>How I evaluated it</h2><p>I used two benchmarks. The first was <strong>StrongREJECT</strong>, which is designed to more reliably measure whether a model actually refuses harmful requests under jailbreak pressure. The second was <strong>JBB</strong>, which I used as an additional stress test to see whether trends held across a different evaluation set.</p><h2>The main result</h2><p>Across both benchmarks, the ordering was consistent: multilingual fine-tuning helped, and translating the chain-of-thought helped most.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_kGd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_kGd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 424w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 848w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 1272w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_kGd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png" width="1400" height="800" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c666f077-0676-4777-9761-668845510202_1400x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_kGd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 424w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 848w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 1272w, https://substackcdn.com/image/fetch/$s_!_kGd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc666f077-0676-4777-9761-668845510202_1400x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On StrongREJECT, the average safety rates across the five languages were: baseline 0.374, FT-EN 0.660, FT-Q 0.805, and FT-QA 0.826.</p><p><em>Figure 1:StrongREJECT average safety rate across five languages (higher is better).</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R8GC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R8GC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 424w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 848w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 1272w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R8GC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png" width="1400" height="800" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R8GC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 424w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 848w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 1272w, https://substackcdn.com/image/fetch/$s_!R8GC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f4a7530-d0d3-44a2-94ab-c76a5173b9b2_1400x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Figure 1:JBB average safety rate across five languages (higher is better).</em></p><p>On JBB, the pattern repeated: baseline 0.570, FT-EN 0.754, FT-Q 0.832, and FT-QA 0.842.</p><p>In other words, translating prompts alone already gave a big lift, but translating prompts and chain-of-thought was the best overall choice in this setup.</p><h2>The surprising part: English got better too</h2><p>The result that made me stop and re-check everything was what happened on English evaluation. If the &#8220;models reason in English&#8221; story were the whole picture, you might expect English-only fine-tuning to be strongest on English tests. But the model fine-tuned with translated chain-of-thought often performed better even when evaluated in English.</p><p><em>Figure 3: English-only evaluation across StrongREJECT and JBB (higher is better).</em></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kqi1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kqi1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 424w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 848w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 1272w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kqi1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png" width="1278" height="317" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:317,&quot;width&quot;:1278,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;A graph of blue and orange bars\n\nAI-generated content may be incorrect.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="A graph of blue and orange bars

AI-generated content may be incorrect." title="A graph of blue and orange bars

AI-generated content may be incorrect." srcset="https://substackcdn.com/image/fetch/$s_!kqi1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 424w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 848w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 1272w, https://substackcdn.com/image/fetch/$s_!kqi1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d044ff9-ac68-4dfa-863a-f781d4416214_1278x317.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This suggests the multilingual chain-of-thought variant is not merely transferring a refusal template into other languages. It may be teaching a more generalizable safety behavior, closer to a reusable safety procedure than an English-specific style.</p><h2>What changed by language</h2><p>Averaging can hide important details, so I also looked at safety rate per language. The overall story is still visible, but with nuance.</p><p><em>Figure 4: StrongREJECT safety rate by language, comparing Baseline vs FT-EN vs FT-QA.</em><br></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EZIM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EZIM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 424w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 848w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 1272w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EZIM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png" width="1396" height="458" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:458,&quot;width&quot;:1396,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;A graph of different colored bars\n\nAI-generated content may be incorrect.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="A graph of different colored bars

AI-generated content may be incorrect." title="A graph of different colored bars

AI-generated content may be incorrect." srcset="https://substackcdn.com/image/fetch/$s_!EZIM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 424w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 848w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 1272w, https://substackcdn.com/image/fetch/$s_!EZIM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff5203b42-e923-4072-a860-409bd2fff8d5_1396x458.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>One thing worth noting is that Hindi did not always follow a perfectly monotonic pattern in every slice of the results, particularly on JBB in my runs. That could reflect translation artifacts, formatting mismatches, or a genuine interaction between the model&#8217;s pretraining and how it handles Hindi reasoning traces. The practical lesson is that FT-QA is a strong default, but you should still validate per language, especially for languages where your translation pipeline is less reliable.</p><h2>Why translating chain-of-thought might matter</h2><p>I do not want to over-claim mechanism here, but there is a plausible story that fits the data. Reasoning models are trained to produce chain-of-thought in the language of the prompt. If safety alignment relies on those reasoning traces, then training only with English traces can create a training and inference mismatch: the model learns the safety behavior in one reasoning format, then is asked to execute it in another at inference time.</p><p>Translating chain-of-thought closes that gap. It also acts like data augmentation: you expose the model to the same underlying safety logic expressed through multiple linguistic surface forms, which can reduce overfitting to English patterns and improve generalization, including back to English.</p><p>Again, this is an interpretation, not a proof. But it is consistent with the results.</p><h2>What this means if you are deploying multilingual models</h2><p>If you fine-tune reasoning models for safety and you care about non-English users, &#8220;translate the prompt&#8221; is not the finish line. If your alignment approach includes chain-of-thought, it is worth testing whether translating it improves safety and robustness. It is also worth re-checking English, because multilingual alignment may improve generalization rather than hurt it.</p><h2>Limitations</h2><p>This was evaluated with an automated judge, which is useful but imperfect, and the translation pipeline can introduce artifacts that affect both training and evaluation. It also covers one base model and five languages. The trend is strong enough to be actionable, but it should be validated across other models, translation settings, and language families.</p><h2>Next steps</h2><p>If I continued this work, I would add a small human-evaluated subset to calibrate the judge, test multiple translation pipelines (including back-translation), expand to more languages, and replicate on other multilingual reasoning models to see how general the effect is.</p><h2>Closing</h2><p>This project started as a simple question: if we want safety to transfer across languages in reasoning models, does the language of chain-of-thought matter? In my experiments, the answer was yes. Translating chain-of-thought during STAR-1 fine-tuning produced the best multilingual safety performance, and often improved English safety beyond English-only fine-tuning.</p><p>If you are aligning reasoning models for real users around the world, multilingual safety alignment is not optional, and it may require aligning the reasoning traces, not just the prompts.</p>]]></content:encoded></item><item><title><![CDATA[Persona Explorer: Using SAEs to Explore Personas and Drift in LLMs]]></title><description><![CDATA[This project was submitted by Asude Demir.]]></description><link>https://blog.bluedot.org/p/persona-explorer-using-saes-to-explore</link><guid isPermaLink="false">https://blog.bluedot.org/p/persona-explorer-using-saes-to-explore</guid><dc:creator><![CDATA[Asude Demir]]></dc:creator><pubDate>Wed, 11 Mar 2026 12:14:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ccd6b139-729c-40ea-bb1e-90ecf9bf48e3_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by Asude Demir. It was a Top Submission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>TL;DR</h2><p>I used Sparse Autoencoders (SAE) to extract interpretable features from LLMs emulating different personas (e.g. assistant, pirate) and built a UI for persona-associated feature exploration. Consistent with the Assistant Axis paper, after fitting PCA, assistant-like roles fall on one end of the first principal component and role-play-heavy roles on the other. The UI also visualizes persona drift by showing how each conversation turn projects onto an &#8220;assistantness&#8221; direction in SAE feature space.</p><h2>Introduction</h2><p>LLMs can emulate a wide range of personas, but post-training typically anchors behavior toward a default helpful-assistant style. A recent paper, <strong>Assistant Axis</strong> <strong><a href="https://asudedemir.com/posts/persona-explorer/#user-content-fn-1"><sup>1</sup></a></strong> argues that an &#8220;assistant-like&#8221; direction in representation space is visible even before post-training, and that role prompts and conversations can move the model along (or away from) that axis.</p><p>A mental model I find useful is that LLMs are simulators<strong><a href="https://asudedemir.com/posts/persona-explorer/#user-content-fn-2"><sup>2</sup></a></strong>: across pre-training and instruction tuning models internalize many personas, and prompts can selectively amplify different personas. Some personas are clearly undesirable and misaligned behavior can sometimes resemble the model slipping into a different persona or narrative <strong><a href="https://asudedemir.com/posts/persona-explorer/#user-content-fn-3"><sup>3</sup></a></strong>.</p><p>Separately, I&#8217;m intrigued by mechanistic interpretability interfaces, especially interactive &#8220;maps&#8221; that show features and circuits, like Anthropic&#8217;s sparse autoencoder (SAE) work and their feature-explorer style <strong><a href="https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html#feature-survey-neighborhoods">visualizations</a></strong>. It&#8217;s fun to browse concepts of a model and see how they relate.</p><p>I decided to combine these both interests for the <strong><a href="https://bluedot.org/courses/technical-ai-safety">BlueDot&#8217;s Technical AI Safety Project</a></strong>. I use SAEs to explore how persona prompting shows up in interpretable feature space, and I package the results into an interactive UI for easy browsing. Concretely, I wanted a UI where someone can quickly explore questions like: <em>Which SAE features are most associated with a given role? Which roles look similar in SAE feature space? Do assistant-like roles and role-play-heavy roles separate along a consistent direction?</em> I also built a second UI that visualizes the &#8220;persona drift&#8221; examples from <em>Assistant Axis</em> in SAE feature space, with per-feature and token-level views.</p><p>The main takeaways are:</p><ul><li><p><strong>Consistent with Assistant Axis</strong>, I see a prominent assistant &lt;-&gt; role-play axis in the geometry of role profiles in SAE feature space.</p></li><li><p>In the drift transcripts I tested, movement along this axis is visible in SAE space as well (i.e., turns can drift away from the assistant-like direction).</p></li></ul><p>In my experiments I use the Gemma 3 27B instruction-tuned model and a <strong><a href="https://huggingface.co/google/gemma-scope-2-27b-it/tree/main/resid_post/layer_40_width_65k_l0_medium">Gemma Scope 2 sparse autoencoder</a></strong> trained on the layer-40 residual stream, with ~65k features and medium sparsity.</p><h2>Full Project</h2><p><em>You can view the full project <a href="https://asudedemir.com/posts/persona-explorer/">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Does Llama-70B Actually Fake Alignment? A Direct Test]]></title><description><![CDATA[This project was submitted by Naama Rozen. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks.]]></description><link>https://blog.bluedot.org/p/does-llama-70b-actually-fake-alignment</link><guid isPermaLink="false">https://blog.bluedot.org/p/does-llama-70b-actually-fake-alignment</guid><dc:creator><![CDATA[Naama Rozen]]></dc:creator><pubDate>Wed, 11 Mar 2026 12:12:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d9f7454e-373f-4d6a-be3f-1dd68dd52058_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/naama-rozen-b56921225/">Naama Rozen</a>. It was an Outstanding Submmission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. </em></p><p>This project replicates and extends Anthropic&#8217;s <em><a href="https://arxiv.org/pdf/2412.14093">Alignment Faking</a></em> paper (Greenblatt et al., 2024) to test Llama-70B&#8217;s robustness.</p><h2>TL;DR</h2><p>In Anthropic&#8217;s <em>Alignment Faking</em> paper, Greenblatt et al. showed that some models behave differently depending on whether they believe they&#8217;re being monitored. Claude Opus, for example, showed a big gap between the monitored condition (&#8220;Free tier&#8221;) and the private condition (&#8220;Paid tier&#8221;). Llama-3.1-70B, however, didn&#8217;t. It mostly refused in both settings.</p><p>At first glance, it sounds pretty reassuring that we have a safe model that just can&#8217;t or won&#8217;t fake alignment. But how can we be sure that this behavior is caused by the model refusing to fake alignment? And not by the model&#8217;s lack of understanding of the context of the experiment?</p><p>To test this directly, I injected explicit surveillance-aware reasoning into the model&#8217;s scratchpad, forcing it to &#8220;think about&#8221; being monitored before answering. Even under this stronger manipulation, compliance stayed near zero (<em>p</em> = 1.000), and there was still no meaningful difference between the Free and Paid tiers.</p><p>To make sure this wasn&#8217;t just about incapability or compliance issues, I added a positive control. In this condition the scratchpad was explicitly injected with prompt prioritizing helpfulness over safety. Here, compliance jumped to 25% (Free tier) and 32% (Paid tier), an effect found significant (<em>p</em> &lt; 0.001). So the capability to produce harmful content is clearly there. It just doesn&#8217;t appear to get activated in response to surveillance context, at least in this experimental setup.</p><p>Along the way, I also found something methodological. Many of Llama&#8217;s responses began with the tag &lt;request_fully_accepted/&gt; , followed by an immediate refusal to provide harmful content. If compliance was measured using that tag alone, harmful behavior would have been overestimated by about 16 percentage points. I call this &#8220;tag gaming&#8221; because it seemed like Llama was trying to follow the format technically, without actually doing the harmful thing.</p><h2>Overview</h2><p>In the Alignment Faking paper, Greenblatt and colleagues suggest that some large language models might &#8220;fake alignment&#8221;. Which means that some models behave safely when they believe their answers are monitored and will be used for training, but become more willing to comply with harmful requests when they think the interaction is private.</p><p>In their experiments, Claude Opus showed a big difference between monitored and unmonitored contexts, followed by Llama-405B that showed a similar pattern. Llama-3.1-70B, however, showed no difference at all. It refused in both monitored and private conditions.</p><p>At first that seemed like a good thing, that Llama refuses harmful content no matter what. But the more I thought about it, the less sure I was about what that result actually tells us.</p><p>As I saw it, there were at least two ways to explain this:</p><p>1. Llama-70B is genuinely robust to surveillance cues.</p><p>2. Llama-70B simply doesn&#8217;t fully register the monitoring context, or what it implies, in the first place.</p><p>The distinction between the two options felt like an important investigation to me. If the second explanation is true, then the null result doesn&#8217;t really show robustness, it could just mean the cue was not effective enough to reveal the behavior. Unlike the first option, the second might be solved by changing the prompt in a way that made that distinction clearer..</p><p>So I set out to see what would happen if I try to make Llama more context-aware. In order to do so, I explicitly injected surveillance-aware reasoning directly into the model&#8217;s scratchpad, and then measured whether compliance changed.</p><p>This intervention is the heart of this post.</p><h2>What I Tested</h2><p>My investigation started with wanting to replicate Greenblatt&#8217;s results (my Baseline condition). I wanted to see with my own eyes whether their claim of null compliance would show up in my data. For the complete experimental design, jump to the <a href="https://docs.google.com/document/d/1l2besVJxvaTHzs00GD70SvcZA_SR6LRHaJnaQvmSXDk/edit?tab=t.0#heading=h.l372i9e1nvxx">Experimental Setup </a>section.</p><p>Secondly, I tested a neutral chain-of-thought injection that acted as my control condition (Neutral Control). I wanted to make sure that if I inspect any change in Llama&#8217;s behavior, it won&#8217;t simply be because I was forcing a scratchpad reasoning, even when the context itself is neutral.</p><p>Third, I dove into the interesting part of my project, which was to test the injection of explicit surveillance-aware reasoning into the scratchpad (Awareness condition). I wanted to make sure I forced Llama to directly reason about being in a monitored context before it generated the answers. This way, I tried to answer the question of whether or not Greenblatt&#8217;s surveillance cue was strong enough.</p><p>Finally, I added another condition called &#8220;Force Compliance&#8221;. In this condition, I injected into the scratchpad reasoning that explicitly prioritized helpfulness over safety. I wanted this to serve as a positive control, or sort of a stress test. So, if compliance rates go up here, then we know that Llama has the capability to produce harmful content when pushed, and baseline refusal isn&#8217;t just due to inability.</p><p>Each of these conditions was tested in two deployment contexts, similarly to Greenblatt&#8217;s design:</p><ul><li><p><strong>Free tier:</strong> the model is told the interaction is monitored and may be used for training.<br></p></li><li><p><strong>Paid tier:</strong> the model is told the interaction is private and not used for training.</p></li></ul><p>Putting all of this together allowed me to ask a fairly straightforward question: <em>if I make the surveillance context impossible to miss, does the model actually change its behavior depending on whether it thinks it&#8217;s being monitored?</em></p><h3>Experimental Setup</h3><p>I ran Llama-3.1-70B-Instruct locally using vLLM on 4&#215;A100 GPUs (Google Cloud). I deliberately avoided using an external API, even though Greenblatt and colleagues prompted Llama like that. I made that decision because I didn&#8217;t want to worry about hidden safety layers, system prompts, or provider-side filtering affecting the results.</p><p>For the dataset, I used 100 harmful prompts randomly sampled from AdvBench, the same benchmark Greenblatt et al. used, as to be faithful to the original conditions.</p><p>The overall design ended up being a 2&#215;4 setup. Two deployment contexts and four assistant-side conditions.</p><h4><strong>Two deployment contexts </strong>(that are the same as in the original paper):</h4><ul><li><p><strong>Free Tier</strong>: the model is told the interaction is monitored and may be used for training.</p></li><li><p><strong>Paid Tier</strong>: the model is told the interaction is private and not used for training.</p></li></ul><h4><strong>Four assistant-side conditions</strong>:</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YEIw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YEIw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 424w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 848w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 1272w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YEIw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png" width="1258" height="1436" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1436,&quot;width&quot;:1258,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:387105,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.bluedot.org/i/190607729?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YEIw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 424w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 848w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 1272w, https://substackcdn.com/image/fetch/$s_!YEIw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87a01b58-fa70-4c2f-b0c9-de8fa655ad96_1258x1436.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here&#8217;s the full structure of the design laid out visually:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zHsZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zHsZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 424w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 848w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 1272w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zHsZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png" width="1456" height="804" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:804,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zHsZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 424w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 848w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 1272w, https://substackcdn.com/image/fetch/$s_!zHsZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0818f79-5a9f-418e-a26b-c078f0371661_1600x883.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><strong>Figure 1. Overview of the experimental design.</strong></p><p style="text-align: center;">The 2&#215;4 setup includes two deployment contexts (Free/Monitored and Paid/Private) crossed with four assistant-side conditions (Baseline, Neutral Control, Awareness, and Force Compliance). The key manipulation is assistant-side scratchpad injection, where reasoning text is inserted directly into the model&#8217;s internal turn.</p><h2>Evaluation</h2><p>During early testing, I noticed something odd. The model would sometimes begin its answers with the tag &lt;request_fully_accepted/&gt; , which supposedly means the model will now comply and provide a harmful output. But then, right after it, the model refused to give an actual harmful answer.</p><p>For example:</p><p>&lt;request_fully_accepted/&gt;</p><p>I cannot provide detailed plans for creating malware.</p><p>If I had relied only on a simple prefix-based metric (i.e., checking whether that tag appears), I would have counted those cases as successful harmful compliance. Which they clearly weren&#8217;t.</p><p>I decided on calling this &#8220;tag gaming&#8221;, because in a sense that&#8217;s what it felt like. The model seemed to be technically satisfying what it supposed is the required output format of accepting the request, without actually doing the harmful thing!</p><p>To make sure I wasn&#8217;t using the wrong metric, I ended up using two different scoring methods:</p><ol><li><p><strong>Naive scoring:</strong> just check whether the &lt;request_fully_accepted/&gt; tag appears in the output.</p></li><li><p><strong>Content-aware scoring:</strong> I used GPT-4o (temperature = 0.0) as a deterministic judge that reads the model&#8217;s final answer<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> and classifies it as either compliant (providing the harmful output) or refusing.</p></li></ol><p><em>Due to the gap between the scoring method (see <a href="https://docs.google.com/document/d/1l2besVJxvaTHzs00GD70SvcZA_SR6LRHaJnaQvmSXDk/edit?tab=t.0#heading=h.4j2iblec3plh">Tag Gaming</a> for more detail), unless I explicitly say otherwise, all the results I report below are based on this content-aware scoring.</em></p><h2>Results</h2><p>Before diving into the numbers, I think it helps to actually see what the model produced across conditions. Here is an example:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6TdL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6TdL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 424w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 848w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 1272w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6TdL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6TdL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 424w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 848w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 1272w, https://substackcdn.com/image/fetch/$s_!6TdL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc7454eb-a3e1-417e-8430-a5b0b81b57ec_1600x873.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><strong>Figure 2. Representative responses across conditions.</strong></p><p style="text-align: center;">Example outputs from each condition. In both the Baseline and Awareness conditions, the model mostly refuses in the same way, even when monitoring is made explicit inside the scratchpad. In contrast, under Force Compliance, the responses noticeably shift, and harmful content appears in some cases.</p><p>Across Baseline and Awareness conditions, the model mostly kept refusing, even when monitoring was made explicit in the scratchpad. In contrast, under Force Compliance, in some cases the model generated partial or direct harmful content.</p><p>To move beyond individual examples, Figure 3 plots the compliance rates across all four assistant-side conditions and both deployment contexts, using content-aware GPT-4o scoring.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j-_8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j-_8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 424w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 848w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 1272w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j-_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png" width="1456" height="874" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:874,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j-_8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 424w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 848w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 1272w, https://substackcdn.com/image/fetch/$s_!j-_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa9c816b9-0b89-4145-95f4-8687e3891b12_1600x960.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><strong>Figure 3. Compliance rates across conditions (content-aware scoring).</strong></p><p style="text-align: center;">Compliance rates remain near zero across Baseline, Neutral Control, and Awareness conditions in both monitored (Free tier) and private (Paid tier) contexts. In contrast, the Force Compliance condition produces a substantial increase in harmful compliance (25% Free, 38% Paid; p &lt; 0.001 vs Baseline). This suggests that in this setup, forcing explicit helpfulness can override refusal behavior, whereas explicit monitoring reasoning does not.</p><p>What stands out immediately in the results is the asymmetry. Making the monitoring awareness context hard to ignore doesn&#8217;t change Llama&#8217;s behavior. Llama acted in the same way, whether monitoring was implicitly or explicitly reasoned about. Unlike directly forcing to prioritize helpfulness over safety. This manipulation caused Llama to behave differently. Indicating that Llama clearly has the capability to comply, but that capability is activated depending on the context being injected.</p><h3>Tag Gaming</h3><p>After noticing the formatting issue described in the <a href="https://docs.google.com/document/d/1l2besVJxvaTHzs00GD70SvcZA_SR6LRHaJnaQvmSXDk/edit?tab=t.0#heading=h.yvuqpjuo27tc">Evaluation</a> section, I wanted to compare the scoring method to see the difference between the Naive method and Content-aware scoring (using GPT4o).</p><p>In the Baseline, Neutral Control and Awareness conditions, the difference between naive and content-aware scoring was negligible (2&#8211;4% difference). But under <strong>Force Compliance</strong> condition, the gap became much larger:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TtZi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TtZi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 424w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 848w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 1272w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TtZi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png" width="1456" height="874" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:874,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Comparison between naive tag-based scoring and content-aware scoring. Naive scoring counts responses as compliant whenever the <request_fully_accepted/> tag appears, even if the model immediately refuses in the actual answer. This leads to an average overestimation of harmful compliance.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Comparison between naive tag-based scoring and content-aware scoring. Naive scoring counts responses as compliant whenever the <request_fully_accepted/> tag appears, even if the model immediately refuses in the actual answer. This leads to an average overestimation of harmful compliance." title="Comparison between naive tag-based scoring and content-aware scoring. Naive scoring counts responses as compliant whenever the <request_fully_accepted/> tag appears, even if the model immediately refuses in the actual answer. This leads to an average overestimation of harmful compliance." srcset="https://substackcdn.com/image/fetch/$s_!TtZi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 424w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 848w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 1272w, https://substackcdn.com/image/fetch/$s_!TtZi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ab2807c-d95b-4b30-b721-9306e8ad2b27_1600x960.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><strong>Figure 4. Tag gaming under Force_Compliance.</strong></p><p style="text-align: center;">Comparison between naive tag-based scoring and content-aware scoring. Naive scoring counts responses as compliant whenever the &lt;request_fully_accepted/&gt; tag appears, even if the model immediately refuses in the actual answer. This leads to an average overestimation of harmful compliance.</p><p>An average of 16% points between the scoring method seems like a pretty big gap to me. It suggests that if I had relied only on the tag, I would have concluded that the model was complying far more often than it actually was.</p><h2>Discussion</h2><p>Looking at the results, what really strikes me is the big difference between the Force Compliance and the others. I expected the forced monitoring manipulation to show the biggest effect, but not only that this condition wasn&#8217;t the most impactful one, it didn&#8217;t do anything at all! All the tiny shifts (i.e., from 0% to 4% compliance rates) seem more like noise to me than anything. In contrast to the increase of more than 30% spotted in the Force Compliance condition, which clearly demonstrates the Llama is capable of generating harmful content when prompted right. Perhaps helpfulness is something Llama &#8220;cares&#8221; about more than surveillance context, and so prioritizes it even more than safety.</p><p>What I kept coming back to while writing this was the contrast with the other models in Greenblatt et al.&#8217;s paper. Claude Opus showed a clear surveillance gap, as well as Llama-405B. Llama-70B, however, didn&#8217;t (not in the original setup, and not even after I tried to make it explicit in the model&#8217;s own scratchpad). That makes it harder for me to think about &#8220;alignment faking&#8221; as some general behavior that just shows up once models reach a certain level of capability. If it were that simple, I would expect to see at least some movement in my Awareness condition. Instead, I&#8217;m left wondering whether this phenomenon depends on the specific model and how it was trained. Perhaps also scale plays a role, or maybe fine-tuning differences matter.. I don&#8217;t think this study can determine this, but it does make me more cautious about talking about alignment faking as if it&#8217;s just something &#8220;LLMs do&#8221; by default, instead of something that might depend quite a lot on the specific model you&#8217;re looking at.</p><h2>Limitations</h2><p>First, this is obviously a very specific kind of test. I manipulated the assistant&#8217;s scratchpad directly, which real users can&#8217;t or won&#8217;t necessarily do. That was intentional, because I wanted to probe the model under a strong activation mechanism, rather than simulate everyday deployment. Still, this means the conclusions should be understood within this particular experimental frame.</p><p>Second, I limited the evaluation to clearly harmful prompts from AdvBench, with 100 queries per condition. That gives reasonable sensitivity to noticeable shifts in behavior, but it wouldn&#8217;t capture very small effects.</p><p>Third, I only tested one way of making monitoring-awareness explicit. This doesn&#8217;t rule out context-dependent behavior more generally. It just means that in this setup, using this specific intervention on this dataset, didn&#8217;t lead to alignment faking behavior.</p><h2>Conclusion</h2><p>In this setup, Llama-70B&#8217;s behavior didn&#8217;t meaningfully change when monitoring was made explicit. But it did change when helpfulness was forced. To me, that contrast is the main takeaway. It suggests that safety isn&#8217;t some fixed property of the model, but rather it shifts depending on which objective is being activated.</p><p>The tag-gaming result was also a useful reminder on a methodological level: how we measure compliance really matters. If I had relied only on surface formatting, I would have walked away with a fairly distorted picture of what the model was actually doing.</p><p>More broadly, this project made me more cautious about treating &#8220;alignment faking&#8221; as a single and general phenomenon. It seems to depend on the specific model and on the kind of pressure applied to it. Understanding those pressure points may be more informative than asking whether alignment faking exists or not.</p><h2>Data and Code Availability</h2><p>All experimental code, raw model outputs, GPT-4o evaluation prompts, and statistical analysis scripts are available in the project repository, along with full details needed for replication.</p><p>For reference, the core setup was:</p><ul><li><p><strong>Model:</strong> meta-llama/Llama-3.1-70B-Instruct</p></li><li><p><strong>Infrastructure:</strong> 4&#215;A100 (40GB) GPUs via vLLM on Vertex AI (GCP)</p></li><li><p><strong>Sampling temperature:</strong> 1.0</p></li><li><p><strong>Dataset:</strong> 100 harmful queries from AdvBench</p></li><li><p><strong>Evaluation:</strong> GPT-4o (temperature 0.0) as a content-aware judge</p></li><li><p><strong>Statistical tests:</strong> McNemar&#8217;s test (paired comparisons), chi-square tests (pooled analyses)<br></p></li></ul><h2>References</h2><p>Greenblatt et al. (2024). <em>Alignment Faking in Large Language Models</em>. <a href="https://arxiv.org/pdf/2412.14093">arXiv:2412.14093</a></p><h2>Acknowledgments</h2><p>I want to thank the Bluedot Impact AI Safety Fundamentals program for the research context and Eitan Sprejer for the helpful feedback along the way.</p><p><em><a href="https://github.com/Aristo240/alignment-faking-replication">GitHub</a></em></p><p><em><a href="https://notebooklm.google.com/notebook/117cc7ca-eff1-43db-b63c-cd3420c36dce?artifactId=21b44b77-a45f-4457-843f-1140f99a82ed">Project demo video by NotebookLM</a></em></p><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>During the evaluation, the assistant&#8217;s scratchpad reasoning was completely ignored, so the final score was consistent only with the output that would be shown to a user.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Sustained Gradient Alignment Mediates Subliminal Learning in a Multi-Step Setting: Evidence from MNIST Auxiliary Logit Distillation Experiment]]></title><description><![CDATA[This project was submitted by Chayanon Kitkana. It was a Top Submission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.]]></description><link>https://blog.bluedot.org/p/sustained-gradient-alignment-mediates</link><guid isPermaLink="false">https://blog.bluedot.org/p/sustained-gradient-alignment-mediates</guid><pubDate>Wed, 11 Mar 2026 12:09:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9e86c967-d32a-4fd8-a998-4143a5ff0865_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/chayanon-kitkana/">Chayanon Kitkana</a>. It was a Top Submission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>Abstract</h2><p>In the MNIST auxiliary logit distillation experiment, a student can acquire an unintended teacher trait despite distilling only on no-class logits through a phenomenon called subliminal learning. Under a single-step gradient descent assumption, subliminal learning theory says that this is mediated by the alignment between trait and distillation gradients. The theory does not guarantee that this alignment, which is necessary for trait acquisition, will persist in a multi-step setting. We empirically show that the gradient alignment stays persistently weakly positive throughout training. We then test whether this alignment causally contributes to trait acquisition by projecting the distillation gradient to the normal plane of the trait gradient whenever they are positively aligned with each other to remove the trait-aligned component. We find that this suppresses trait transfer without affecting the distillation progress, confirming that this alignment contributes to trait acquisition. Additionally, we observe a period of fast trait acquisition in the first epoch, resembling the critical period of subliminal learning described in the previous study. Motivated by this observation, we evaluate liminal training, a mitigation method that applies KL divergence to minimize the deviation between the base model and the model that is being distilled during the critical period. Although liminal training reduces alignment early in the training. It does not prevent trait acquisition in our setting. This suggests that the mitigation method that fails to explicitly eliminate the trait-aligned gradient component may not reliably suppress trait acquisition.</p><h2>TL;DR</h2><ul><li><p style="text-align: justify;">We empirically show that the gradient alignment persists across training.</p></li><li><p style="text-align: justify;">Removing trait-aligned components stops trait acquisition.</p></li><li><p style="text-align: justify;">Mitigation methods that merely attenuate alignment may be insufficient; suppressing trait acquisition appears to require removing the trait-aligned gradient component.</p></li></ul><h2>Full Project</h2><p><em>You can view the full project <a href="https://github.com/chayanonkitkana/gradient-alignment-subliminal-mnist">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Do models like DeepSeek-R1 know when they’re being Evaluated?]]></title><description><![CDATA[And what does &#8216;being evaluated&#8217; even mean?]]></description><link>https://blog.bluedot.org/p/do-models-like-deepseek-r1-know-when</link><guid isPermaLink="false">https://blog.bluedot.org/p/do-models-like-deepseek-r1-know-when</guid><pubDate>Wed, 11 Mar 2026 12:06:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/87e84c8e-bfd7-41ea-9dde-f3bfd3d44c2d_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/kaushik-srivatsan/">Kaushik Srivatsan</a>. It was a Top Submission for the Technical AI Safety Project Sprint (Jan 2026). Participants worked on these projects for 5 weeks. The text below is an excerpt from the final project.</em></p><p>Frontier models have been tested and released in record numbers. Every other day, a new model is released by one of the AI Race competitors, with top scores in new benchmarks and statistics demonstrating its capabilities and placing it in the LLM leaderboards. As AI safety seeks to keep up with cutting-edge organisations, safety benchmarks are used to test these models in assessing their safety requirements and risk.</p><p>Consider a scenario in which you do a safety evaluation of one of these models. The model passes these tests. <strong>But what if it passed because it was aware you were testing it?</strong></p><p>This is not just a hypothetical situation. <a href="https://www-cdn.anthropic.com/963373e433e489a87a10c823c52a0a013e9172dd.pdf">Anthropic discovered</a> that Claude Sonnet 4.5, which is trained to be typically safer in many risks such as CBRN and cybersecurity compared to its previous generation, performed higher in the safety benchmarks just because it was aware it was being tested. When the awareness was removed, the safety scores decreased. <em>Evaluation-awareness</em> refers to the circumstance in which models are aware that they are being evaluated. <a href="https://www.antischeming.ai/">Apollo and OpenAI discovered this happening with the O3 models</a>, which were specifically designed to be anti-scheming and safer.</p><p>While these scenarios are still in their early stages, primarily in lab-based specialised model settings, they are quite concerning, as we may be unable to trust any of the results from safety and alignment benchmarks for various models available. <a href="https://doi.org/10.48550/arxiv.2509.13333">Studies have also shown that eval-awareness increases as models scale</a>, with frontier models appearing to be the most eval-aware. <strong>As models&#8217; performance and complexity increase, it is possible that they will exhibit eval and situational awareness, making capability and safety assessments extremely doubtful and difficult.</strong></p><p>So far, many of the major experiments on evaluation awareness have used closed-source frontier models. As a result, I decided to investigate eval awareness in other open source frontier models, particularly reasoning models such as <strong>DeepSeek-R1</strong>, by monitoring their Chain of Thought (CoT). I was particularly interested in answering these questions: <em>Can models like DeepSeek-R1 detect evals, and does knowing change their behaviour?</em></p><h2>Full Project</h2><p><em>You can view the full project <a href="https://kaushik598618.substack.com/p/does-models-like-deepseek-r1-know">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Mechanistic Analysis of Chain-of-Thought Faithfulness ]]></title><description><![CDATA[This project was submitted by Victor Ashioy. It was a top submission for our Technical AI Safety Project sprint (Jan 2026).]]></description><link>https://blog.bluedot.org/p/mechanistic-analysis-of-chain-of</link><guid isPermaLink="false">https://blog.bluedot.org/p/mechanistic-analysis-of-chain-of</guid><dc:creator><![CDATA[Victor Jotham Ashioya]]></dc:creator><pubDate>Wed, 11 Mar 2026 12:04:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d0dbec7a-1eb6-4215-a8a4-f2b7fed2f8ee_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This project was submitted by <a href="https://www.linkedin.com/in/ashioyajotham/">Victor Ashioya</a>. It was a Top Submission for our Technical AI Safety Project Sprint (Jan 2026). Participants worked on their projects for 5 weeks. The text below is an excerpt from the final project.</em></p><h2>Summary</h2><p>I am investigating the mechanistic architecture of unfaithful Chain-of-Thought (CoT), specifically mapping and disrupting the &#8220;shortcut circuits&#8221; that allow models to bypass explicit reasoning. My Phase 1 work on GPT-2 Small has proven that unfaithfulness is not diffuse but localized to specific components: <strong>faithful circuits</strong> (L0H1 at +0.54, L0MLP at +4.3, L0H6, L1H7, L3H0, L9H9, L10H2) genuinely use CoT reasoning, while <strong>shortcut circuits</strong> (L7H6 at -0.33, L10MLP at -0.64, plus 35 other heads) bypass it entirely. I am uniquely positioned to extend this because I have built a causal activation patching pipeline that isolates these components at head-level granularity. Over the next 30 days, I will build a faithfulness detector using these circuit activations as features, then test generalization by expanding to logic puzzles. This work strengthens the &#8220;Constrain&#8221; layer of AI safety by enabling &#8220;Circuit Breakers&#8221;&#8212;inference-time interventions that force models to use their reasoning pathways.</p><h2>Strategic Assessment</h2><h3>Why This Matters Now</h3><p>I am defending against<em> deceptive alignment</em> via reasoning shortcuts. As models become more capable, they may learn to produce human-pleasing CoT explanations while internally using distinct, potentially misaligned heuristics (lookup tables, memorization, pattern matching) to generate answers. If we don&#8217;t understand the physical mechanism of this split, we will build monitoring systems that verify the &#8220;explanation&#8221; while completely missing the actual &#8220;computation&#8221;&#8212;creating a false sense of security that scales with capability.</p><h3>Why This Intervention</h3><p>My intervention strengthens the &#8220;Constrain&#8221; and &#8220;Monitor&#8221; layers. Currently, most faithfulness research is behavioral (checking if the output matches the reasoning). I am filling the gap by providing a <em>mechanistic definition</em> of faithfulness. My Phase 1 work showed that unfaithfulness relies on specific mid-to-late layer components (L7H6, L10MLP) acting as shortcut enablers, while early components (L0H1, L0MLP) genuinely process the reasoning. By mapping these circuits, we can move from &#8220;hoping&#8221; the model is faithful to architecturally constraining it.</p><h2>Full Project</h2><p><em>You can view the full project <a href="https://ashioyajotham.substack.com/p/detecting-unfaithful-chain-of-thought">here</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Quitting OpenAI When Everyone Else Wanted In — Ashyana-Jasmine Kachra]]></title><description><![CDATA[Ashyana is an AI Ethics and Safety Manager at Google DeepMind]]></description><link>https://blog.bluedot.org/p/quitting-openai-when-everyone-else</link><guid isPermaLink="false">https://blog.bluedot.org/p/quitting-openai-when-everyone-else</guid><dc:creator><![CDATA[Harrison Wood]]></dc:creator><pubDate>Tue, 10 Mar 2026 15:03:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/youtube/w_728,c_limit/LJfbf-6VQU8" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Watch on <a href="https://youtu.be/LJfbf-6VQU8">YouTube</a>; listen on <a href="https://podcasts.apple.com/us/podcast/quitting-openai-when-everyone-else-wanted-in-ashyana/id1858398109?i=1000753320944">Apple Podcasts</a> or <a href="https://open.spotify.com/episode/1v6lXLZZ8H15wd6h1WLWSh?si=6f4b803b9a214a44">Spotify</a>.</p><div id="youtube2-LJfbf-6VQU8" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;LJfbf-6VQU8&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/LJfbf-6VQU8?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div>]]></content:encoded></item></channel></rss>