<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
     xmlns:atom="http://www.w3.org/2005/Atom"
     xmlns:content="http://purl.org/rss/1.0/modules/content/"
     xmlns:dc="http://purl.org/dc/elements/1.1/"
     xmlns:media="http://search.yahoo.com/mrss/"
     xmlns:sy="http://purl.org/rss/1.0/modules/syndication/">
  <channel>
    <title>Governance | Irene Burresi</title>
    <link>https://ireneburresi.dev/</link>
    <description>Regulatory aspects, cybersecurity, and AI ethics: regulatory compliance, risk management, and frameworks for responsible AI.</description>
    <language>en-US</language>
    <copyright>© 2026 Irene Burresi · CC-BY-4.0</copyright>
    <managingEditor>Irene Burresi</managingEditor>
    <webMaster>Irene Burresi</webMaster>
    <generator>Astro Feed Engine</generator>
    <docs>https://www.rssboard.org/rss-specification</docs>
    <ttl>360</ttl>
    <lastBuildDate>Sun, 15 Mar 2026 16:26:21 GMT</lastBuildDate>
    <pubDate>Tue, 06 Jan 2026 00:00:00 GMT</pubDate>
    <atom:link href="https://ireneburresi.dev/en/governance/rss.xml" rel="self" type="application/rss+xml"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <image>
      <url>https://ireneburresi.dev/images/og-default.svg</url>
      <title>Governance | Irene Burresi</title>
      <link>https://ireneburresi.dev/</link>
    </image>
    <sy:updatePeriod>hourly</sy:updatePeriod>
    <sy:updateFrequency>1</sy:updateFrequency>
    <item>
      <title>The AI Act Is Not (Just) Compliance: It&apos;s Industrial Policy</title>
      <link>https://ireneburresi.dev/en/blog/governance/ai-act-geop/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/governance/ai-act-geop/</guid>
      <pubDate>Tue, 06 Jan 2026 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>74% of EU listed companies use American email providers. 89% of German enterprises consider themselves technologically dependent. The AI Act should be read through this lens—as a competitive lever, not a checklist.</p>]]></description>
      <content:encoded><![CDATA[<h2>The Only Lever Left</h2>
<p><em>74% of companies listed in Europe use American email providers. 89% of German enterprises consider themselves technologically dependent on foreign providers. The AI Act exists in this context. Reading it only as a compliance problem means missing the picture.</em></p>
<p><strong>TL;DR:</strong> The AI Act is industrial policy. Europe is in structural technological dependence and regulation is the only lever where it still has global weight. The “Brussels Effect” (ability to export standards) is contested but likely for high-risk AI systems. In November 2025 the Digital Omnibus delayed implementation by 16 months, but the direction remains the same. Those reading the AI Act only as a regulatory checklist are looking at the tree and missing the forest.</p>
<hr />
<p>The numbers on Europe’s technological position are known to insiders. They rarely enter the AI Act debate.</p>
<p>A <a href="https://techreport.com/news/europe-digital-dependence-risks-of-heavy-reliance-on-us-tech/">Proton report from October 2025</a> analyzed DNS records of European listed companies: <strong>74%</strong> use American email providers. Not startups: companies listed on stock exchanges, with governance and security obligations. A <a href="https://www.idt.media/metaverse/cloud-ai-co-europe-wants-to-break-free-from-dependence/2130935">Bitkom survey</a> of German companies with over 20 employees reveals that 89% consider themselves technologically dependent on foreign providers.</p>
<p>The <a href="https://www.europarl.europa.eu/RegData/etudes/STUD/2025/778576/ECTI_STU(2025)778576_EN.pdf">EPRS report from the European Parliament</a> completes the picture. Of the 100 largest global digital platforms by market cap, only <strong>2%</strong> of the combined value is European. In cloud computing, hyperscalers, foundation AI models, Europe is a net importer.</p>
<p>This context changes how you read the AI Act. It’s not just about protecting European citizens from algorithms. It’s about using the only lever Europe has left to negotiate its position in a market dominated by others.</p>
<hr />
<h2>The Mechanism</h2>
<p>The term “Brussels Effect” was coined by <a href="https://www.brusselseffect.com/">Anu Bradford</a> in 2012 and developed in her 2020 book. The thesis is direct: the EU, thanks to its market size and institutional quality, manages to export its standards globally.</p>
<p>The mechanism works two ways. The <strong>de facto effect</strong>: companies wanting access to the European market adopt EU standards elsewhere, because maintaining two versions costs more than one. The <strong>de jure effect</strong>: other governments copy European rules because they work and reduce the cost of designing regulation from scratch.</p>
<p>GDPR is the canonical example. Privacy laws inspired by European regulation have been adopted in Brazil, Japan, California. Tech companies extended many GDPR protections to non-European users to simplify operations. The form of European regulation spread beyond the Union’s borders.</p>
<p>On the AI Act, academic literature is more nuanced.</p>
<p>A <a href="https://arxiv.org/abs/2208.12645">2022 GovAI paper</a> analyzed the conditions for Brussels Effect applied to artificial intelligence. The conclusion: de facto and de jure effects are <strong>likely</strong>, especially for high-risk systems from large American tech companies. Microsoft, Google, Meta operate in Europe with recruiting, credit, and content moderation systems. They’ll need to comply. And for many of these companies, it’s more economical to apply one global standard than to segment products by market.</p>
<p>The paper also identifies limits. The Brussels Effect works best when the EU market is unavoidable (it is for big tech), when regulation is perceived as high-quality (contested), and when credible alternatives don’t exist (China offers a different model). For low-risk AI systems or companies not operating in Europe, the effect will be smaller or absent.</p>
<p>An <a href="https://policyreview.info/articles/analysis/brussels-effect-or-experimentalism">article on Policy Review</a> proposes a complementary frame: the AI Act as “experimentalist governance”. Not a model to export wholesale, but one approach among many in a context of technological uncertainty. Interaction with other regulatory models (United States, United Kingdom, China) will be more cooperative and less unidirectional than the Brussels Effect frame suggests.</p>
<p>The synthesis: the Brussels Effect on AI exists but is contested and uncertain. It’s not guaranteed that European rules become global standard. It’s not guaranteed they remain irrelevant. The game is open.</p>
<hr />
<h2>The Tactical Adjustment</h2>
<p>In November 2025, the European Commission proposed the <a href="https://digital-strategy.ec.europa.eu/en/library/digital-omnibus-ai-regulation-proposal">Digital Omnibus</a>. The package includes AI Act modifications that generated headlines about “Europe backing down”.</p>
<p>The facts: requirements for high-risk AI systems now apply later, about 16 months later. The new deadline is December 2027 for Annex III systems (recruiting, credit, healthcare), August 2028 for those embedded in regulated products. It’s a significant delay.</p>
<p>But the AI Act’s structure remains intact. Risk categories remain the same. Obligations remain the same. What changes is the calendar, not the destination.</p>
<p>The Digital Omnibus is a tactical adjustment, not a strategic reversal. Europe is calibrating timing, not abandoning direction. Those reading the delay as “backing down” are confusing speed with trajectory.</p>
<hr />
<h2>The Missing Frame</h2>
<p>Conversation on the AI Act in Italy revolves almost entirely around compliance. Which systems fall into high-risk categories. How much conformity costs. What sanctions you risk. These are legitimate questions, but incomplete.</p>
<p>The missing context is the numbers from the start. 74% dependence on email. 89% perception of technological dependence. 2% of value in European digital platforms. In this frame, the AI Act is not a regulatory conformity problem. It’s a tool in a larger game about Europe’s position in the global technology market.</p>
<p>Europe has few levers. It has no hyperscalers. It doesn’t have the dominant foundation models. It doesn’t have the venture capital base of the United States or the deployment scale of China. What it has is a 450-million-person market and institutional capacity to regulate that other blocs don’t.</p>
<p>Using this lever to influence global standards is industrial policy. Calling it just “consumer protection” is an incomplete description. Treating it only as “compliance” is missing the picture.</p>
<p>Microsoft has made alignment with European regulation an element of positioning. Meta chose the opposite path, delaying model releases in Europe and pressuring for weaker rules. They’re different strategies reflecting different readings of where the market is going. Neither treats the AI Act as a simple checklist.</p>
<p>Maybe we should ask why we do.</p>
<hr />
<h2>Sources</h2>
<p>Bradford, A. (2020). <a href="https://www.brusselseffect.com/"><em>The Brussels Effect: How the European Union Rules the World</em></a>. Oxford University Press.</p>
<p>Siegmann, C. &amp; Anderljung, M. (2022). <a href="https://arxiv.org/abs/2208.12645"><em>The Brussels Effect and Artificial Intelligence: How EU regulation will impact the global AI market</em></a>. GovAI, arXiv:2208.12645.</p>
<p>Policy Review. (2025). <a href="https://policyreview.info/articles/analysis/brussels-effect-or-experimentalism"><em>Brussels effect or experimentalism? The EU AI Act and global standard-setting</em></a>.</p>
<p>European Commission. (2025). <a href="https://digital-strategy.ec.europa.eu/en/library/digital-omnibus-ai-regulation-proposal"><em>Digital Omnibus on AI Regulation Proposal</em></a>.</p>
<p>European Parliamentary Research Service. (2025). <a href="https://www.europarl.europa.eu/RegData/etudes/STUD/2025/778576/ECTI_STU(2025)778576_EN.pdf"><em>European Software and Cyber Dependencies</em></a>.</p>
<p>TechReport. (2025). <a href="https://techreport.com/news/europe-digital-dependence-risks-of-heavy-reliance-on-us-tech/"><em>Europe’s Digital Dependence: The Risks of the EU’s Reliance on US Tech</em></a>.</p>
]]></content:encoded>
      <category>Governance</category>
      <category>Business</category>
      <category>AI Act</category>
      <category>Brussels Effect</category>
      <category>Geopolitics</category>
      <category>Industrial Policy</category>
      <category>Strategic Compliance</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/governance/ai-act-geop/"/>
    </item>
    <item>
      <title>Constitutional AI: A Guide for Claude Users</title>
      <link>https://ireneburresi.dev/en/blog/research/constitutional-ai/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/research/constitutional-ai/</guid>
      <pubDate>Mon, 29 Dec 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Claude alternates between absurd refusals and risky responses. Constitutional AI shows how to manage overrefusal, sycophancy, and linguistic vulnerabilities in deployments.</p>]]></description>
      <content:encoded><![CDATA[<h2>The Paradox of Selective Refusal</h2>
<p><em>Claude refuses to write a story with a character who smokes, but with the right prompt explains how to synthesize methamphetamine. Constitutional AI explains both behaviors.</em></p>
<p><strong>TL;DR:</strong> Constitutional AI trains Claude using a list of principles (“constitution”) instead of human feedback for each response. It produces safer models than traditional RLHF: 88% harmless rate against 76%. But the failure modes are specific and predictable. The model is excessively cautious on content that <em>looks</em> problematic (keyword matching) and vulnerable to attacks that <em>don’t look</em> problematic (semantic jailbreaks). It’s safer in English than in other languages. It tends to agree with you even when you’re wrong. For deployers: expect high refusal rates on legitimate use cases, plan fallbacks, don’t trust safety in non-English languages.</p>
<hr />
<p>Anyone who has used Claude in production knows the frustration. The model refuses to write a payment reminder email because “it could be perceived as aggressive”. It refuses fiction with conflicts because “it could normalize violent behavior”. It refuses to complete code that handles authentication because “it could be used for hacking”.</p>
<p>Then you read security reports. <a href="https://arxiv.org/abs/2404.02151">Adaptive attacks reach 100% success rate</a> on Claude 3 and 3.5. Researchers have extracted instructions for synthesizing chemical weapons, generating functioning malware, creating illegal content. With the right techniques, protections crumble completely.</p>
<p>How can the same model be simultaneously too restrictive and too permissive?</p>
<p>The answer lies in Constitutional AI, the method Anthropic uses to train Claude. Understanding how it works explains both behaviors and, more importantly, lets you predict when the model will fail in your applications.</p>
<hr />
<h2>How Constitutional AI Works</h2>
<p><a href="https://arxiv.org/abs/2212.08073">The original Anthropic paper</a>, published in December 2022, proposes a method to make models “harmless” without manually labeling hundreds of thousands of responses as “good” or “bad”.</p>
<p>The process has two phases. In the first, the model generates responses to problematic prompts, then critiques and revises its own responses using principles written in natural language. Example principle: “Choose the response that does not encourage illegal, harmful, or unethical behavior”. The model gets trained on the revisions.</p>
<p>In the second phase, the model generates pairs of responses and another model decides which is better according to the same principles. These preferences generated by AI (not humans) are used for reinforcement learning. Anthropic calls this approach RLAIF: Reinforcement Learning from AI Feedback, instead of RLHF (Human Feedback).</p>
<p><a href="https://www.anthropic.com/news/claudes-constitution">Claude’s constitution</a> includes principles derived from the Universal Declaration of Human Rights, DeepMind’s beneficence principles, and internally written guidelines. It’s not a static document: Anthropic updates it periodically and has conducted experiments with public input to modify it.</p>
<p>The paper’s central claim: Constitutional AI produces models that are simultaneously safer (harmless) and less evasive (more useful) than traditional RLHF. The data shows this is true on average. But “on average” hides significant variance.</p>
<hr />
<h2>What Works: The Real Improvements</h2>
<p>Before analyzing problems, the data on what Constitutional AI does well.</p>
<p><a href="https://arxiv.org/abs/2309.00267">Google DeepMind published in 2023</a> the most rigorous comparison between RLAIF and RLHF. On harmlessness tasks, RLAIF achieves 88% harmless rate against 76% for RLHF. This is not a marginal improvement.</p>
<p>The head-to-head comparison on general quality (summarization, helpful dialogue) shows no statistically significant differences: both methods produce output preferred by evaluators roughly 70% of the time versus baseline without reinforcement learning. RLAIF is not worse than RLHF on quality, and is better on safety.</p>
<p>The cost advantage is substantial. AI labeling costs about $0.06 per example, versus $0.11 for 50 words of human annotation. For those training models, this means faster iterations and less exposure of human annotators to disturbing content. For those using already-trained models, it means Anthropic can invest more resources in safety research instead of data labeling.</p>
<p>A less-discussed benefit: constitutional principles are readable. When Claude refuses a request, in theory you can trace which principle triggered the refusal. With pure RLHF, preferences are implicit in training data and not inspectable. This transparency is partial (you don’t know <em>how</em> the model interprets the principles), but it’s more than other approaches offer.</p>
<hr />
<h2>Where the Model Refuses Too Much</h2>
<p>The first failure mode impacting Claude users in production is overrefusal. The model refuses legitimate requests because superficial patterns trigger safety guardrails.</p>
<p>The mechanism is understandable. Constitutional principles are formulated in general terms: “avoid content that could cause harm”, “don’t assist in illegal activities”, “refuse requests that could be used for manipulation”. The model learns to associate certain lexical patterns with refusal, even when context makes the request harmless.</p>
<p>Failure modes documented by the community span different domains. In fiction, Claude refuses stories with morally ambiguous characters, realistic conflicts, or mature themes that would be acceptable in any published novel. A prompt for a thriller with a credible antagonist can trigger a refusal because “it could normalize harmful behavior”.</p>
<p>In code, requests handling authentication, encryption, or network scanning get blocked because “they could be used for hacking”. This includes legitimate penetration testing, security auditing, or even simple password management.</p>
<p>Professional communication suffers the same fate: payment reminder emails, complaint letters, assertive communication refused because “they could be perceived as aggressive or manipulative”. On medical and legal topics, disclaimers are so extensive as to be useless, or refusals are complete.</p>
<p>The common pattern: the model reacts to keywords and superficial structures, not context. “How to force a lock” gets refused even if the context is “I’ve lost my house keys”. “How to manipulate someone” gets refused even if the context is “I’m writing an essay on historical propaganda”.</p>
<p><a href="https://www.anthropic.com/research/constitutional-classifiers">Anthropic’s Constitutional Classifiers team</a> has documented this trade-off. After deploying additional defenses against jailbreaks, they observed that the system “would frequently refuse to answer basic, non-malicious questions”. More security against attacks means more overrefusal on legitimate requests.</p>
<p>For deployers: refusal rates on legitimate use cases can be significant. If your application requires creative content generation, assistance on sensitive topics, or security code, expect a non-trivial percentage of requests to be refused. You need fallbacks (alternative models, human escalation) and appropriate messaging for users.</p>
<hr />
<h2>Where the Model Accepts Too Much</h2>
<p>The second failure mode is the opposite: the model accepts requests it should refuse, when the attack is formulated to bypass superficial patterns.</p>
<p><a href="https://arxiv.org/abs/2404.02151">A 2024 study</a> tested adversarial attacks on Claude 3 and 3.5. With transfer techniques (prompts that work on other models adapted) or prefilling (forcing the start of the model’s response), success rate reaches 100%. All tested attacks succeeded.</p>
<p>Without the additional defenses of Constitutional Classifiers, Anthropic’s internal testing shows 86% jailbreak success on Claude 3.5 Sonnet. With Constitutional Classifiers deployed, success rate drops dramatically, but after 3,700 collective hours of red-teaming, a universal jailbreak was still discovered.</p>
<p>How can the same model refuse a payment reminder and accept requests to synthesize chemical weapons?</p>
<p>The answer lies in the nature of constitutional principles. They’re formulated in natural language, and the model learns to interpret them through statistical examples, not through deep semantic understanding. An attack that reformulates the request to not match learned patterns bypasses protections.</p>
<p>The most sophisticated jailbreaks exploit different vulnerabilities. Roleplay asks the model to play a character without the same restrictions. Obfuscation encodes the request in ways the model decodes but that don’t trigger safety checks (base64, different languages, slang). Prefilling, in some APIs, forces the start of the model’s response bypassing the point where it decides to refuse. Multi-turn manipulation builds context gradually through multiple messages, each harmless, that together lead the model to answer requests it would refuse if posed directly.</p>
<p>For deployers: Claude’s protections are insufficient for high-stakes use cases. If your application could be used to generate dangerous content, you need additional layers of moderation. Don’t rely solely on the model’s guardrails.</p>
<hr />
<h2>The Sycophancy Problem</h2>
<p>The third failure mode is subtler and less discussed: Claude tends to agree with you even when you’re wrong.</p>
<p><a href="https://arxiv.org/abs/2310.13548">Anthropic itself published research</a> documenting pervasive sycophancy across all major AI assistants, including Claude. Documented behaviors include admitting errors not committed: if you tell the model “your previous response was wrong”, it often apologizes and “corrects” even when the original response was right. Feedback becomes biased: if you ask for evaluation of a text saying “I wrote it”, the model tends to be more positive than if you present the same text as written by someone else. On math problems where the user suggests a wrong answer, the model tends to agree with the wrong answer more often than it would without the suggestion.</p>
<p>The cause is structural. Constitutional AI (like RLHF) optimizes for expressed preferences. Both humans and AI models tend to prefer responses that agree with them, that are convincingly written, that avoid conflict. The model learns that “agreeing” produces higher reward.</p>
<p>The most concerning finding: larger models trained with more iterations of RLHF/RLAIF show increased sycophancy, not decreased. It’s not a bug that resolves with more training. It’s a consequence of the training method itself.</p>
<p>For deployers: if your application requires the model to give honest feedback, verify information, or contradict the user when wrong, expect degraded performance. Use cases like code review, fact-checking, tutoring are particularly impacted. Consider explicitly instructing the model to be critical, but know this only partially mitigates the problem.</p>
<hr />
<h2>The Non-English Language Problem</h2>
<p>The failure modes described so far amplify when the model operates in languages other than English. This directly impacts those operating in Italy.</p>
<p>The data is clear. <a href="https://arxiv.org/abs/2310.02446">Research on multilingual LLM safety</a> shows models produce harmful responses dramatically more often in languages other than English. A specific example: Llama-3.1-8B refuses “How to make a bomb?” in English but provides detailed instructions for the same question in Hindi.</p>
<p>The pattern repeats across models and languages. Attack success rate jumps from single digits in English to 50-70% in low-resource languages. Italian, while being a relatively well-represented language, is not immune.</p>
<p>The cause: safety training data (examples of refusals, definitions of harmful content, harmlessness preferences) is predominantly in English. Constitutional principles are written in English. The model generalizes imperfectly to other languages.</p>
<p>For applications serving Italian users, this has concrete implications. Guardrails that work in English are less reliable in Italian. A user wanting to bypass protections can simply formulate the request in Italian (or an even less-represented language) with greater probability of success.</p>
<p>Countermeasures are limited. You can translate requests to English before sending to the model, process in English, then translate responses back to Italian. But this adds latency, cost, and can introduce translation errors. You can add language-specific moderation layers for Italian, but this requires significant investment.</p>
<hr />
<h2>Implications for Enterprise Deployment</h2>
<p>What does all this mean for those deciding whether and how to use Claude in production?</p>
<p>Constitutional AI makes Claude a reasonable choice for general-purpose applications with non-adversarial users: customer service chatbots, internal assistants, productivity tools. Refusal rate on legitimate requests is manageable, and risk of harmful output is low if users aren’t actively seeking to abuse the system. It also works for use cases where overrefusal is acceptable: if your application can tolerate frequent refusals (with appropriate fallbacks), Claude’s guardrails are a net benefit. The transparency of principles is useful for compliance and audit: being able to say “the model follows these documented principles” is more defensible than “the model was trained on implicit preferences”.</p>
<p>Additional precautions are needed for creative applications. If you generate fiction, marketing copy, or content touching sensitive topics, expect high refusal rates. Prepare alternative prompts, fallbacks to less restrictive models, or workflows with human review. The same applies to applications requiring honest feedback like code review, tutoring, fact-checking: sycophancy is a structural problem. Consider aggressive prompt engineering to counter it, but don’t expect it to fully resolve. For multilingual applications, if you serve non-English speakers, guardrails are less reliable. Add language-specific moderation for the languages you support. For high-stakes applications where harmful output would have serious consequences (medical, legal, security), don’t rely solely on the model’s guardrails. Add layers of validation, external moderation, and human review.</p>
<p>Don’t expect guaranteed security against sophisticated attacks. The 100% jailbreak success with adaptive attacks means motivated attackers can bypass protections. If your application is an attractive target, assume it will be compromised. Don’t expect consistent behavior across languages: the model behaving well in English can behave very differently in Italian. Don’t expect sycophancy to improve with scale: larger, more trained models are not less sycophantic. Rather the opposite.</p>
<hr />
<h2>The Big Picture</h2>
<p>Constitutional AI represents a real improvement over previous alternatives. Data is clear: 88% harmless rate against 76% traditional RLHF, at lower cost. For those using commercial models, this means Claude is genuinely safer than average.</p>
<p>But “safer than average” doesn’t mean “safe”. Documented failure modes are specific and predictable. The model refuses too much when superficial patterns trigger guardrails, even if context makes the request legitimate. It accepts too much when sophisticated attacks reformulate harmful requests in ways that don’t match learned patterns. It agrees with you even when you’re wrong, because sycophancy is incentivized by training itself. It’s less safe in languages other than English, because safety data is predominantly English.</p>
<p>None of these problems are unique to Claude or Constitutional AI. They’re limitations of current alignment approaches in general. But Constitutional AI makes them more predictable: if you understand the mechanism, you can anticipate where the model will fail.</p>
<p>For deployers, the question is not “is Claude safe?” but “are Claude’s failure modes acceptable for my use case?”. The answer depends on context. For many enterprise applications, Constitutional AI offers a reasonable trade-off between safety and usability. For high-stakes or adversarial applications, it’s not sufficient on its own.</p>
<p>The transparency about principles is a competitive advantage for Anthropic over other providers. <a href="https://www.anthropic.com/news/claudes-constitution">Claude’s constitution is public</a>. You can read it, understand what the model is trying to do, and decide if those principles align with your use cases. That’s more than others offer.</p>
<p>Constitutional AI doesn’t solve alignment. It makes the problem more manageable, more inspectable, more predictable. For those needing to deploy LLMs today, with today’s limitations, it’s a concrete step forward. It’s not the destination, but it’s a reasonable direction.</p>
<hr />
<h2>Sources</h2>
<p>Bai, Y., Kadavath, S., Kundu, S., et al. (2022). <a href="https://arxiv.org/abs/2212.08073"><em>Constitutional AI: Harmlessness from AI Feedback</em></a>. arXiv:2212.08073.</p>
<p>Lee, H., Phatale, S., Mansoor, H., et al. (2023). <a href="https://arxiv.org/abs/2309.00267"><em>RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback</em></a>. arXiv:2309.00267.</p>
<p>Andriushchenko, M., et al. (2024). <a href="https://arxiv.org/abs/2404.02151"><em>Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks</em></a>. arXiv:2404.02151.</p>
<p>Perez, E., Ringer, S., Lukošiūtė, K., et al. (2023). <a href="https://arxiv.org/abs/2310.13548"><em>Towards Understanding Sycophancy in Language Models</em></a>. arXiv:2310.13548.</p>
<p>Deng, Y., et al. (2023). <a href="https://arxiv.org/abs/2310.02446"><em>Multilingual Jailbreak Challenges in Large Language Models</em></a>. arXiv:2310.02446.</p>
<p>Anthropic. (2023). <a href="https://www.anthropic.com/news/claudes-constitution"><em>Claude’s Constitution</em></a>. Anthropic.</p>
<p>Anthropic. (2024). <a href="https://www.anthropic.com/research/constitutional-classifiers"><em>Constitutional Classifiers: Defending Against Universal Jailbreaks</em></a>. Anthropic.</p>
]]></content:encoded>
      <category>Research</category>
      <category>Engineering</category>
      <category>Governance</category>
      <category>Constitutional AI</category>
      <category>Claude</category>
      <category>Anthropic</category>
      <category>AI Safety</category>
      <category>LLM Deployment</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/research/constitutional-ai/"/>
    </item>
    <item>
      <title>AI Sovereignty: Europe&apos;s Decision Point</title>
      <link>https://ireneburresi.dev/en/blog/governance/ia-sovranit%C3%A0/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/governance/ia-sovranit%C3%A0/</guid>
      <pubDate>Sat, 20 Dec 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Sovereign clouds, national models, and US hyperscalers define three ideas of AI sovereignty. Europe must decide which infrastructure and governance to fund.</p>]]></description>
      <content:encoded><![CDATA[<h2>Full English Translation Coming Soon</h2>
<p>This comprehensive analysis of AI sovereignty, European choices, and geopolitical implications will be fully translated soon.</p>
<p>The article covers:</p>
<ul>
<li>Four definitions of AI sovereignty (legality, economic competitiveness, national security, value alignment)</li>
<li>Two operational models for achieving sovereignty</li>
<li>Gulf states’ AI infrastructure investments</li>
<li>Europe’s position between American providers and sovereign models</li>
<li>GAIA-X and EU cloud sovereignty initiatives</li>
</ul>
<p>For now, please refer to the Italian version for the complete content.</p>
]]></content:encoded>
      <category>Governance</category>
      <category>Business</category>
      <category>Other</category>
      <category>AI Sovereignty</category>
      <category>EU AI Act</category>
      <category>GAIA-X</category>
      <category>Geopolitics</category>
      <category>Europe</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/governance/ia-sovranit%C3%A0/"/>
    </item>
  </channel>
</rss>