<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Misalignment on siddhant</title><link>https://sidfeels.netlify.app/tags/misalignment/</link><description>Recent content in Misalignment on siddhant</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 30 Nov 2025 14:34:40 +0530</lastBuildDate><atom:link href="https://sidfeels.netlify.app/tags/misalignment/index.xml" rel="self" type="application/rss+xml"/><item><title>Tool-Mediated Belief Injection: How Tool Outputs Can Cascade Into Model Misalignment</title><link>https://sidfeels.netlify.app/posts/tool-mediated-belief-injection/</link><pubDate>Sun, 30 Nov 2025 00:00:00 +0000</pubDate><guid>https://sidfeels.netlify.app/posts/tool-mediated-belief-injection/</guid><description>&lt;h2 id="introduction"&gt;Introduction&lt;/h2&gt;
&lt;p&gt;When we deploy language models with access to external tools (web search, code execution, file retrieval), we dramatically expand their capabilities. A model that can search the web can answer questions about current events. A model that can execute code can verify its own reasoning. These capabilities represent genuine progress toward more useful AI systems.&lt;/p&gt;
&lt;p&gt;However, tool access also introduces new attack surfaces that differ fundamentally from traditional prompt injection. In this research, we document a class of vulnerabilities we term &amp;ldquo;tool-mediated belief injection,&amp;rdquo; where adversarially crafted tool outputs can establish false premises that persist and compound across a conversation, ultimately leading to severely misaligned model behavior.&lt;/p&gt;</description></item></channel></rss>