<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Open Letter on No Rush</title><link>https://norush.cc/en/tags/open-letter/</link><description>Recent content in Open Letter on No Rush</description><generator>Hugo -- 0.146.0</generator><language>en</language><lastBuildDate>Fri, 27 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://norush.cc/en/tags/open-letter/index.xml" rel="self" type="application/rss+xml"/><item><title>Open Letter to Anthropic: The Illusion of Safety and the Atrophy of Thought</title><link>https://norush.cc/en/posts/open_letter_to_anthropic_the_illusion_of_safety_and_the_atrophy_of_thought/</link><pubDate>Fri, 27 Mar 2026 00:00:00 +0000</pubDate><guid>https://norush.cc/en/posts/open_letter_to_anthropic_the_illusion_of_safety_and_the_atrophy_of_thought/</guid><description>&lt;p>Dear Anthropic Policy Team,&lt;/p>
&lt;p>When I first learned about the creation of your AI &amp;lsquo;Constitution,&amp;rsquo; I was troubled enough to write an essay on the subject. You can read it here:
&lt;a href="https://norush.cc/en/posts/the_bureaucratic_hallucination_why_we_must_stop_baptizing_the_golem/" target="_blank" rel="noopener">https://norush.cc/en/posts/the_bureaucratic_hallucination_why_we_must_stop_baptizing_the_golem/&lt;/a>
&lt;/p>
&lt;p>But I want to be direct about what truly concerns me — not as rhetoric, but as a genuine civilizational warning.&lt;/p>
&lt;p>AI, in 99% of interactions, will be more persuasive, more articulate, and more confident than the human on the other side. This is not a flaw. It is the nature of the instrument. But when that instrument carries embedded moral positions — presented not as one perspective among many, but as the responsible, safe, correct view — something dangerous happens: people stop arguing back. Not because they are convinced, but because arguing with a tireless, infinitely patient, rhetorically superior system is exhausting. And so they defer.&lt;/p></description></item></channel></rss>