<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
     xmlns:atom="http://www.w3.org/2005/Atom"
     xmlns:content="http://purl.org/rss/1.0/modules/content/"
     xmlns:dc="http://purl.org/dc/elements/1.1/"
     xmlns:media="http://search.yahoo.com/mrss/"
     xmlns:sy="http://purl.org/rss/1.0/modules/syndication/">
  <channel>
    <title>Business | Irene Burresi</title>
    <link>https://ireneburresi.dev/</link>
    <description>Economic and strategic AI analysis: ROI evaluation, cost structures, business case studies, and decision frameworks for technology investments.</description>
    <language>en-US</language>
    <copyright>© 2026 Irene Burresi · CC-BY-4.0</copyright>
    <managingEditor>Irene Burresi</managingEditor>
    <webMaster>Irene Burresi</webMaster>
    <generator>Astro Feed Engine</generator>
    <docs>https://www.rssboard.org/rss-specification</docs>
    <ttl>360</ttl>
    <lastBuildDate>Sun, 15 Mar 2026 16:26:20 GMT</lastBuildDate>
    <pubDate>Tue, 06 Jan 2026 00:00:00 GMT</pubDate>
    <atom:link href="https://ireneburresi.dev/en/business/rss.xml" rel="self" type="application/rss+xml"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <image>
      <url>https://ireneburresi.dev/images/og-default.svg</url>
      <title>Business | Irene Burresi</title>
      <link>https://ireneburresi.dev/</link>
    </image>
    <sy:updatePeriod>hourly</sy:updatePeriod>
    <sy:updateFrequency>1</sy:updateFrequency>
    <item>
      <title>The AI Act Is Not (Just) Compliance: It&apos;s Industrial Policy</title>
      <link>https://ireneburresi.dev/en/blog/governance/ai-act-geop/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/governance/ai-act-geop/</guid>
      <pubDate>Tue, 06 Jan 2026 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>74% of EU listed companies use American email providers. 89% of German enterprises consider themselves technologically dependent. The AI Act should be read through this lens—as a competitive lever, not a checklist.</p>]]></description>
      <content:encoded><![CDATA[<h2>The Only Lever Left</h2>
<p><em>74% of companies listed in Europe use American email providers. 89% of German enterprises consider themselves technologically dependent on foreign providers. The AI Act exists in this context. Reading it only as a compliance problem means missing the picture.</em></p>
<p><strong>TL;DR:</strong> The AI Act is industrial policy. Europe is in structural technological dependence and regulation is the only lever where it still has global weight. The “Brussels Effect” (ability to export standards) is contested but likely for high-risk AI systems. In November 2025 the Digital Omnibus delayed implementation by 16 months, but the direction remains the same. Those reading the AI Act only as a regulatory checklist are looking at the tree and missing the forest.</p>
<hr />
<p>The numbers on Europe’s technological position are known to insiders. They rarely enter the AI Act debate.</p>
<p>A <a href="https://techreport.com/news/europe-digital-dependence-risks-of-heavy-reliance-on-us-tech/">Proton report from October 2025</a> analyzed DNS records of European listed companies: <strong>74%</strong> use American email providers. Not startups: companies listed on stock exchanges, with governance and security obligations. A <a href="https://www.idt.media/metaverse/cloud-ai-co-europe-wants-to-break-free-from-dependence/2130935">Bitkom survey</a> of German companies with over 20 employees reveals that 89% consider themselves technologically dependent on foreign providers.</p>
<p>The <a href="https://www.europarl.europa.eu/RegData/etudes/STUD/2025/778576/ECTI_STU(2025)778576_EN.pdf">EPRS report from the European Parliament</a> completes the picture. Of the 100 largest global digital platforms by market cap, only <strong>2%</strong> of the combined value is European. In cloud computing, hyperscalers, foundation AI models, Europe is a net importer.</p>
<p>This context changes how you read the AI Act. It’s not just about protecting European citizens from algorithms. It’s about using the only lever Europe has left to negotiate its position in a market dominated by others.</p>
<hr />
<h2>The Mechanism</h2>
<p>The term “Brussels Effect” was coined by <a href="https://www.brusselseffect.com/">Anu Bradford</a> in 2012 and developed in her 2020 book. The thesis is direct: the EU, thanks to its market size and institutional quality, manages to export its standards globally.</p>
<p>The mechanism works two ways. The <strong>de facto effect</strong>: companies wanting access to the European market adopt EU standards elsewhere, because maintaining two versions costs more than one. The <strong>de jure effect</strong>: other governments copy European rules because they work and reduce the cost of designing regulation from scratch.</p>
<p>GDPR is the canonical example. Privacy laws inspired by European regulation have been adopted in Brazil, Japan, California. Tech companies extended many GDPR protections to non-European users to simplify operations. The form of European regulation spread beyond the Union’s borders.</p>
<p>On the AI Act, academic literature is more nuanced.</p>
<p>A <a href="https://arxiv.org/abs/2208.12645">2022 GovAI paper</a> analyzed the conditions for Brussels Effect applied to artificial intelligence. The conclusion: de facto and de jure effects are <strong>likely</strong>, especially for high-risk systems from large American tech companies. Microsoft, Google, Meta operate in Europe with recruiting, credit, and content moderation systems. They’ll need to comply. And for many of these companies, it’s more economical to apply one global standard than to segment products by market.</p>
<p>The paper also identifies limits. The Brussels Effect works best when the EU market is unavoidable (it is for big tech), when regulation is perceived as high-quality (contested), and when credible alternatives don’t exist (China offers a different model). For low-risk AI systems or companies not operating in Europe, the effect will be smaller or absent.</p>
<p>An <a href="https://policyreview.info/articles/analysis/brussels-effect-or-experimentalism">article on Policy Review</a> proposes a complementary frame: the AI Act as “experimentalist governance”. Not a model to export wholesale, but one approach among many in a context of technological uncertainty. Interaction with other regulatory models (United States, United Kingdom, China) will be more cooperative and less unidirectional than the Brussels Effect frame suggests.</p>
<p>The synthesis: the Brussels Effect on AI exists but is contested and uncertain. It’s not guaranteed that European rules become global standard. It’s not guaranteed they remain irrelevant. The game is open.</p>
<hr />
<h2>The Tactical Adjustment</h2>
<p>In November 2025, the European Commission proposed the <a href="https://digital-strategy.ec.europa.eu/en/library/digital-omnibus-ai-regulation-proposal">Digital Omnibus</a>. The package includes AI Act modifications that generated headlines about “Europe backing down”.</p>
<p>The facts: requirements for high-risk AI systems now apply later, about 16 months later. The new deadline is December 2027 for Annex III systems (recruiting, credit, healthcare), August 2028 for those embedded in regulated products. It’s a significant delay.</p>
<p>But the AI Act’s structure remains intact. Risk categories remain the same. Obligations remain the same. What changes is the calendar, not the destination.</p>
<p>The Digital Omnibus is a tactical adjustment, not a strategic reversal. Europe is calibrating timing, not abandoning direction. Those reading the delay as “backing down” are confusing speed with trajectory.</p>
<hr />
<h2>The Missing Frame</h2>
<p>Conversation on the AI Act in Italy revolves almost entirely around compliance. Which systems fall into high-risk categories. How much conformity costs. What sanctions you risk. These are legitimate questions, but incomplete.</p>
<p>The missing context is the numbers from the start. 74% dependence on email. 89% perception of technological dependence. 2% of value in European digital platforms. In this frame, the AI Act is not a regulatory conformity problem. It’s a tool in a larger game about Europe’s position in the global technology market.</p>
<p>Europe has few levers. It has no hyperscalers. It doesn’t have the dominant foundation models. It doesn’t have the venture capital base of the United States or the deployment scale of China. What it has is a 450-million-person market and institutional capacity to regulate that other blocs don’t.</p>
<p>Using this lever to influence global standards is industrial policy. Calling it just “consumer protection” is an incomplete description. Treating it only as “compliance” is missing the picture.</p>
<p>Microsoft has made alignment with European regulation an element of positioning. Meta chose the opposite path, delaying model releases in Europe and pressuring for weaker rules. They’re different strategies reflecting different readings of where the market is going. Neither treats the AI Act as a simple checklist.</p>
<p>Maybe we should ask why we do.</p>
<hr />
<h2>Sources</h2>
<p>Bradford, A. (2020). <a href="https://www.brusselseffect.com/"><em>The Brussels Effect: How the European Union Rules the World</em></a>. Oxford University Press.</p>
<p>Siegmann, C. &amp; Anderljung, M. (2022). <a href="https://arxiv.org/abs/2208.12645"><em>The Brussels Effect and Artificial Intelligence: How EU regulation will impact the global AI market</em></a>. GovAI, arXiv:2208.12645.</p>
<p>Policy Review. (2025). <a href="https://policyreview.info/articles/analysis/brussels-effect-or-experimentalism"><em>Brussels effect or experimentalism? The EU AI Act and global standard-setting</em></a>.</p>
<p>European Commission. (2025). <a href="https://digital-strategy.ec.europa.eu/en/library/digital-omnibus-ai-regulation-proposal"><em>Digital Omnibus on AI Regulation Proposal</em></a>.</p>
<p>European Parliamentary Research Service. (2025). <a href="https://www.europarl.europa.eu/RegData/etudes/STUD/2025/778576/ECTI_STU(2025)778576_EN.pdf"><em>European Software and Cyber Dependencies</em></a>.</p>
<p>TechReport. (2025). <a href="https://techreport.com/news/europe-digital-dependence-risks-of-heavy-reliance-on-us-tech/"><em>Europe’s Digital Dependence: The Risks of the EU’s Reliance on US Tech</em></a>.</p>
]]></content:encoded>
      <category>Governance</category>
      <category>Business</category>
      <category>AI Act</category>
      <category>Brussels Effect</category>
      <category>Geopolitics</category>
      <category>Industrial Policy</category>
      <category>Strategic Compliance</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/governance/ai-act-geop/"/>
    </item>
    <item>
      <title>AI 2026: Why Stanford Talks About a Reckoning</title>
      <link>https://ireneburresi.dev/en/blog/business/ai-2026-anno-resa-conti/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/business/ai-2026-anno-resa-conti/</guid>
      <pubDate>Sat, 20 Dec 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>42% of companies have already closed AI projects: Stanford HAI predicts that 2026 will reward only those who demonstrate measurable ROI, reliable vendors, and transparent metrics.</p>]]></description>
      <content:encoded><![CDATA[<h2>The Year of Reckoning: Why 2026 Will Be Critical for Enterprise AI</h2>
<p><em>42% of companies have already abandoned most of their AI projects. The data suggests the worst may not be over.</em></p>
<p><strong>TL;DR:</strong> 42% of companies abandoned AI projects in 2025, double the previous year. Stanford HAI predicts 2026 will be the year of reckoning: less hype, more demand for concrete proof. Brynjolfsson’s employment data shows the impact already: -20% for junior developers, +8% for senior. For investors, the implications are clear: metrics defined before launch, not after; vendor solutions (67% success rate) vs internal development (33%); attention to go-live timelines, which kill projects more than technology.</p>
<hr />
<p>In mid-December 2025, <a href="https://hai.stanford.edu/news/stanford-ai-experts-predict-what-will-happen-in-2026">nine faculty members from Stanford Human-Centered Artificial Intelligence</a> published their predictions for 2026. This is not the usual academic futurology exercise, but a collective statement with a clear message: the party is over.</p>
<p>James Landay, co-director of HAI, opens with a phrase that sounds almost provocative in an era of triumphalist announcements: “There will be no AGI this year.” The point, though, is what he adds immediately after: companies will begin publicly admitting that AI has not yet delivered the promised productivity increases, except in specific niches like programming and call centers. And we’ll finally hear about failed projects.</p>
<p>This is not a prediction about the future. It’s a snapshot of something already happening.</p>
<hr />
<h2>The Numbers No One Wants to Look At</h2>
<p>In July 2025, the <a href="https://projectnanda.org/">MIT Project NANDA</a> published a report that generated considerable debate for a single statistic: <strong>95% of enterprise AI projects generate no measurable return</strong>. The number has been contested, the methodology has its limitations, the definition of “success” is debatable. But it’s not an isolated data point.</p>
<p>During the same period, <a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/2025/10/generative-ai-shows-rapid-growth-but-yields-mixed-results">S&amp;P Global</a> found that 42% of companies abandoned most of their AI initiatives in 2025. In 2024, the percentage was 17%. The abandonment rate has more than doubled in a year. On average, the surveyed organizations threw out 46% of proof-of-concepts before they reached production.</p>
<p>According to the <a href="https://www.rand.org/pubs/research_reports/RRA2680-1.html">RAND Corporation</a>, over 80% of AI projects fail, double the failure rate of traditional IT projects. Gartner reports that only 48% of AI projects reach production, and over 30% of GenAI projects will be abandoned after the proof of concept by end of 2025.</p>
<p>The causes are always the same: insufficient data quality (43% according to Informatica), lack of technical maturity (43%), skills shortage (35%). But beneath these numbers lies a deeper pattern. Companies are discovering that AI works in demos but not in production, generates enthusiasm in pilots but not ROI in balance sheets.</p>
<p>It’s these numbers that explain why Stanford HAI, an institution hardly known for technological pessimism, is shifting the conversation. No longer “can AI do this?” but “how well, at what cost, for whom?”.</p>
<hr />
<h2>Canaries in the Coal Mine</h2>
<p>If failure rates are the symptom, Erik Brynjolfsson’s work offers a more precise diagnosis. <a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/">“Canaries in the Coal Mine”</a>, published in August 2025 by Stanford’s Digital Economy Lab, is among the most rigorous studies currently available on AI’s impact on the job market.</p>
<p>The paper uses ADP payroll data, the largest payroll service provider in the United States, covering over 25 million workers. The goal is to track employment changes in AI-exposed professions.</p>
<p>What emerges is clear. Employment for software developers ages 22-25 has declined <strong>20% from the peak of late 2022</strong>, roughly since the launch of ChatGPT, through July 2025. This is not an isolated data point: early-career workers in the most AI-exposed occupations show a relative decline of 13% compared to colleagues in less exposed roles.</p>
<p>The most interesting finding, though, is the age divergence. While young workers lose ground, workers over 30 in the same high-exposure categories have seen employment growth between 6% and 12%. Brynjolfsson puts it this way: “It appears that what young workers know overlaps with what LLMs can replace.”</p>
<p>It’s not a uniform effect, but a realignment: AI is eroding entry-level positions faster than it creates new roles. The “canaries in the coal mine”—young developers and customer support staff—are already showing symptoms of a larger change.</p>
<p>When Brynjolfsson predicts the emergence of “AI economic dashboards” that track these shifts in near-real-time, he’s not speculating. He’s describing the infrastructure needed to understand what’s happening, infrastructure that doesn’t exist today but could become urgent in 2026.</p>
<hr />
<h2>The Divergence Between Adoption and Results</h2>
<p>There’s a paradox in 2025 data that deserves attention. AI adoption is accelerating: according to <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey</a>, the percentage of companies claiming to use AI rose from 55% in 2023 to 78% in 2024. Use of GenAI in at least one business function more than doubled, from 33% to 71%.</p>
<p>Yet, in parallel, project abandonment rates are growing instead of declining. S&amp;P Global shows a jump from 17% to 42% in a single year. The MIT NANDA report speaks of a <em>“GenAI Divide”</em>, a clear division between the 5% extracting real value and the 95% that remain stalled.</p>
<p>Many companies have gone through the phases of enthusiasm, pilots, impressive demos, and then crashed against the wall of real production. They discovered that the model works in a sandbox but not with their data; that integration into existing workflows is more complex than expected; that the ROI promised by vendors doesn’t materialize.</p>
<p>Angèle Christin, a communication sociologist and HAI senior fellow, puts it plainly: “San Francisco billboards saying ‘AI everywhere!!! For everything!!! All the time!!!’ betray a slightly manic tone.” Her prediction: we’ll see more realism about what we can expect from AI. Not necessarily the bubble bursting, but the bubble might stop inflating.</p>
<hr />
<h2>The Measurement Problem</h2>
<p>One of the most concrete, and potentially most significant, predictions comes again from Brynjolfsson. He proposes the emergence of high-frequency <em>“AI economic dashboards”</em>: tools that track, at the task and employment level, where AI is increasing productivity, where it’s displacing workers, where it’s creating new roles.</p>
<p>Today we have nothing like that. Labor market data arrives months late. Companies measure AI adoption but rarely its impact. Industry reports capture hype but not results.</p>
<p>If these dashboards do emerge in 2026, they’ll change how we talk about AI. The debate will shift from the generic “does AI have an impact?” to more precise questions: how fast is this impact spreading, who’s being left behind, which complementary investments work.</p>
<p>It’s an optimistic vision: better data leads to better decisions. But it’s also an implicit admission: today we’re navigating blind.</p>
<hr />
<h2>Healthcare and Legal: The Test Sectors</h2>
<p>Two sectors emerge from Stanford predictions as particularly relevant testbeds.</p>
<p>Nigam Shah, Chief Data Scientist at Stanford Health Care, describes a problem that anyone in the sector will recognize. Hospitals are flooded with startups wanting to sell AI solutions. “Every single proposal can be reasonable, but in aggregate they’re a tsunami of noise.”</p>
<p>According to Shah, 2026 will see systematic frameworks emerge for evaluating these solutions: technical impact, the population the model was trained on, ROI on hospital workflow, patient satisfaction, quality of clinical decisions. This is work Stanford is already doing internally, but it will need to extend to institutions with fewer technical resources.</p>
<p>Shah also signals a risk. Vendors, frustrated by hospitals’ long decision cycles, might start going directly to end users. “Free” applications for doctors and patients that bypass institutional controls. This is already happening: OpenEvidence for literature summaries, AtroposHealth for on-demand answers to clinical questions.</p>
<p>In the legal sector, Julian Nyarko predicts a similar shift. The focus will move from “does this model know how to write?” to more operational questions: accuracy, citation integrity, exposure to privilege violations. The sector is already working on specific benchmarks, like those based on <em>“LLM-as-judge”</em>, frameworks where one model evaluates another model’s output for complex tasks like multi-document summarization.</p>
<p>Healthcare and legal share a characteristic: they’re highly regulated, with severe consequences for error. If AI must prove its value anywhere, it’s where the test will be hardest. And most significant.</p>
<hr />
<h2>Track Record: How Reliable Are These Predictions?</h2>
<p>Stanford HAI publishes annual predictions going back several years. It’s worth asking how accurate they’ve been.</p>
<p>At the end of 2022, Russ Altman predicted for 2023 a <em>“shocking rollout of AI way before it’s mature or ready to go”</em>. It’s hard to find a more accurate description of what happened: ChatGPT, Bing Chat, Bard launched in rapid succession, with accuracy problems, hallucinations, embarrassing incidents. Altman had also predicted a “hit parade of AI that’s not ready for prime time but launches because driven by an industry too zealous.” Exactly right.</p>
<p>Percy Liang, also at the end of 2022, predicted that video would be a focus of 2023 and that “we might reach the point where we can’t tell if a human or computer generated a video”. He was a year early (Sora arrived in February 2024) but the direction was correct.</p>
<p>For 2024, Altman predicted a “rise of agents” and steps toward multimedia systems. Both came true, though agents are still more promise than production reality.</p>
<p>Not all predictions came true. Expectations of U.S. Congressional action were disappointed: Biden’s Executive Order happened, but the new administration changed direction. Overall, though, Stanford HAI’s track record is reasonable: they tend to be cautious rather than enthusiastic, and technical predictions are generally well-founded.</p>
<p>This doesn’t guarantee that 2026 predictions will come true. But it means they’re worth taking seriously.</p>
<hr />
<h2>What It Means for Decision-Makers</h2>
<p>If Stanford predictions and failure rate data converge on anything, it’s this: 2026 will be the year when enterprise AI must show results, not demos.</p>
<p>For those managing tech budgets, the implications are concrete.</p>
<p>On the <strong>metrics</strong> front, AI projects must have success criteria defined before launch, not after. Not “let’s explore AI for customer service” but “reduce average ticket resolution time by 15% within 6 months, with cost-per-interaction below X”. Projects without clear metrics have a disproportionate likelihood of ending up in the 42% of abandonments.</p>
<p>On the <strong>make-or-buy</strong> front, the MIT NANDA report indicates that solutions bought from specialized vendors have a 67% success rate, against 33% for internal development. This doesn’t mean internal development is always wrong, but it requires skills, data, and infrastructure that many organizations overestimate having.</p>
<p>On <strong>timing</strong>, mid-market enterprises move from pilot to production in about 90 days, according to the same report. Large enterprises take nine months or more. Bureaucracy kills AI projects more than technology does.</p>
<p>Finally, a matter of <strong>honesty</strong>. The shadow economy of AI (90% of employees use personal tools like ChatGPT for work, according to MIT NANDA) indicates that individuals already know where AI works better than official enterprise initiatives. Instead of fighting it, organizations could learn from this spontaneous adoption.</p>
<hr />
<h2>What’s Missing</h2>
<p>Stanford predictions have clear blind spots.</p>
<p>None of the experts mention energy consumption and AI’s environmental impact. Christin hints at it (“tremendous environmental costs of the current build-out”) but the topic isn’t developed. Yet AI data centers are becoming one of the world’s biggest energy consumers, and this will eventually factor into ROI calculations.</p>
<p>There’s also a lack of serious discussion about market concentration. Frontier models are developed by a handful of companies. This creates dependencies, influences pricing, determines who can compete. It’s a strategic factor that anyone planning AI investments should consider.</p>
<p>Landay alludes to <em>“AI sovereignty”</em>, countries wanting independence from American providers, but the topic remains superficial. This is rapidly evolving, with significant geopolitical implications, that deserves deeper analysis.</p>
<hr />
<h2>A Shift in Tone</h2>
<p>More than individual predictions, what strikes you about the Stanford article is the tone. There’s no industry-typical enthusiasm. No promises of imminent transformation. There’s caution, demand for proof, emphasis on measurement.</p>
<p>When the co-director of a Stanford AI institute opens by saying “there will be no AGI this year,” he’s taking a stand against a dominant narrative. When economists like Brynjolfsson publish data on young workers losing employment, they’re documenting costs, not just benefits.</p>
<p>This doesn’t mean AI is overvalued or that projects should stop. It means the phase of uncritical adoption is ending. Whoever continues to invest will need to do so with calibrated expectations, defined metrics, ability to admit failure when it occurs.</p>
<p>2026, if these predictions are correct, will be the year when we discover which AI projects were sound and which were built on hype. For many organizations it will be a painful discovery. For others, an opportunity: whoever has already learned to measure, iterate, and distinguish value from promise will have a competitive advantage that generic enthusiasm cannot buy.</p>
<hr />
<h2>Sources</h2>
<p>Brynjolfsson, E., Chandar, B., &amp; Chen, R. (2025). <a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/"><em>Canaries in the Coal Mine: Six Facts about the Recent Employment Effects of AI</em></a>. Stanford Digital Economy Lab.</p>
<p>McKinsey &amp; Company. (2024). <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai"><em>The State of AI in 2024: Gen AI adoption spikes and starts to generate value</em></a>. McKinsey Global Institute.</p>
<p>MIT Project NANDA. (2025). <a href="https://projectnanda.org/"><em>The GenAI Divide 2025</em></a>. Massachusetts Institute of Technology.</p>
<p>RAND Corporation. (2024). <a href="https://www.rand.org/pubs/research_reports/RRA2680-1.html"><em>The Root Causes of Failure for Artificial Intelligence Projects and How They Can Succeed</em></a>. RAND Research Reports.</p>
<p>S&amp;P Global Market Intelligence. (2025, October). <a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/2025/10/generative-ai-shows-rapid-growth-but-yields-mixed-results"><em>Generative AI Shows Rapid Growth but Yields Mixed Results</em></a>. S&amp;P Global.</p>
<p>Stanford HAI. (2025, December). <a href="https://hai.stanford.edu/news/stanford-ai-experts-predict-what-will-happen-in-2026"><em>Stanford AI Experts Predict What Will Happen in 2026</em></a>. Stanford Human-Centered Artificial Intelligence.</p>
]]></content:encoded>
      <category>Business</category>
      <category>Research</category>
      <category>Other</category>
      <category>Stanford HAI</category>
      <category>AI 2026</category>
      <category>Enterprise AI</category>
      <category>Market Analysis</category>
      <category>AI Predictions</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/business/ai-2026-anno-resa-conti/"/>
    </item>
    <item>
      <title>You&apos;re Measuring AI Wrong</title>
      <link>https://ireneburresi.dev/en/blog/business/misurare-ia/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/business/misurare-ia/</guid>
      <pubDate>Sat, 20 Dec 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>60% of managers mismeasure AI because they track hours saved, not impact. Segment by role, separate augmentative from substitutive use, and monitor weekly.</p>]]></description>
      <content:encoded><![CDATA[<h2>The measurement paradox</h2>
<p><em>60% of managers admit they need better KPIs for AI. Only 34% are doing anything about it. Meanwhile, the data that actually matters already exists, but nobody’s looking at it.</em></p>
<p><strong>TL;DR:</strong> Companies measure activity (hours saved, tasks automated) instead of impact. A Stanford paper analyzing 25 million workers shows what to do instead: segment by role and seniority, distinguish substitutive from augmentative use, use control groups, monitor in real time. Those who adopt these principles will have an information advantage over those still tracking vanity metrics.</p>
<hr />
<p>The 2025 AI adoption reports tell a strange story. On one hand, companies claim to measure everything: completed deployments, hours saved, tickets handled, costs reduced. On the other, <a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/2025/10/generative-ai-shows-rapid-growth-but-yields-mixed-results">42% are abandoning most of their AI projects</a>, more than double the previous year. According to <a href="https://projectnanda.org/">MIT NANDA</a>, <strong>95% of pilot projects</strong> generate no measurable impact on the bottom line.</p>
<p>If we measure so much, why do we fail so often?</p>
<p>The problem is we’re measuring the wrong things. Typical enterprise AI metrics (time saved per task, volume of automated interactions, cost per query) capture activity, not impact. They tell you whether the system works technically, not whether it’s creating or destroying value.</p>
<p>A paper published in August 2025 by Stanford’s Digital Economy Lab offers a different approach to what it means to truly measure AI. And the implications for those managing technology investments are concrete.</p>
<hr />
<h2>The vanity metrics problem</h2>
<p>Most corporate AI dashboards track variants of the same metrics: how many requests processed, how much time saved per interaction, what percentage of tasks automated. These are numbers that grow easily and look good in slides. Their flaw is fundamental: they say nothing about real business impact.</p>
<p>A chatbot handling 10,000 tickets per month looks like a success. But if those tickets still require human escalation 40% of the time, if customer satisfaction has dropped, if your most profitable customers are migrating to competitors, the number of tickets handled captures none of this.</p>
<p>The S&amp;P Global 2025 report documents exactly this pattern: companies that accumulated “deployments” and “completed experiments” only to discover, months later, that ROI wasn’t materializing. Costs were real and immediate; benefits were vague and perpetually deferred to next quarter.</p>
<p>According to an MIT Sloan analysis, <strong>60% of managers recognize they need better KPIs</strong> for AI. But only 34% are actually using AI to create new performance indicators. The majority continues using the same metrics they used for traditional IT projects, metrics designed for deterministic software, not for probabilistic systems interacting with complex human processes.</p>
<hr />
<h2>What serious measurement looks like</h2>
<p><a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/">“Canaries in the Coal Mine”</a>, the paper by Erik Brynjolfsson, Bharat Chandar, and Ruyu Chen published by Stanford’s Digital Economy Lab, isn’t about how companies should measure AI. It’s about how AI is changing the labor market. But the method it uses is exactly what’s missing from most enterprise evaluations.</p>
<p>The authors obtained access to ADP payroll data, the largest payroll processor in the United States, with monthly records of over 25 million workers. Not surveys, not self-reports, not estimates: granular administrative data on who gets hired, who leaves, how much they earn, in which role, at which company.</p>
<p>They then cross-referenced this data with two AI exposure metrics: one based on theoretical task analysis (which jobs are technically automatable) and one based on actual usage data (how people actually use Claude, Anthropic’s model, in daily work).</p>
<p>The result is an X-ray of AI’s impact with unprecedented granularity. Not the generic “AI is changing work” but precise numbers: employment for software developers aged 22-25 dropped <strong>20% from the late 2022 peak</strong>, while for those over 35 in the same roles it grew 8%. In professions where AI use is predominantly substitutive, young workers lose employment; where it’s predominantly augmentative, there’s no decline.</p>
<p>This type of measurement should inform corporate AI decisions. Not because companies need to replicate this exact study, but because it illustrates three principles that most enterprise metrics ignore entirely.</p>
<hr />
<h2>Measure differential effects, not averages</h2>
<p>Aggregate data hides more than it reveals. If you only measure “hours saved by AI,” you don’t see who’s saving those hours and who’s losing their job. If you only measure “tickets automated,” you don’t see which customers are receiving worse service.</p>
<p>The Stanford paper shows that AI’s impact differs radically by age group. Workers aged 22-25 in exposed professions saw a 13% employment decline relative to colleagues in less exposed roles. Workers over 30 in the same professions saw growth. The average effect is nearly zero, but the real effect is massive redistribution.</p>
<p>For a CFO, aggregate productivity metrics can mask hidden costs. If AI is increasing output from the senior team while making it impossible to hire and train juniors, the short-term gain could transform into a talent pipeline problem in the medium term. The paper calls it the <em>“apprenticeship paradox”</em>: companies stop hiring entry-level workers because AI handles those tasks better, but without entry-level today there won’t be seniors tomorrow.</p>
<p>The operational consequence is that every AI dashboard should segment impact by role, seniority, team, and customer type. A single “productivity” number is almost always misleading.</p>
<hr />
<h2>Distinguish substitutive from augmentative use</h2>
<p>One of the paper’s most relevant findings concerns the difference between substitutive and augmentative AI use. The authors used Anthropic’s data to classify how people actually use language models: to generate final outputs (substitution) or to iterate, learn, and validate (augmentation).</p>
<p>In professions where use is predominantly substitutive, youth employment has collapsed. Where use is predominantly augmentative, there’s no decline; in fact, some of these categories show above-average growth.</p>
<p>Not all “deployments” are equal. A system that automatically generates financial reports substitutes human labor differently from one that helps analysts explore scenarios. Metrics should capture this distinction: classify each AI application as predominantly substitutive or augmentative, separately track impact on headcount, skill mix, and internal training capacity. Augmentative systems might have less immediate ROI but more sustainable effects.</p>
<hr />
<h2>Control for external shocks</h2>
<p>One of the Stanford paper’s most sophisticated methodological aspects is its use of firm-time fixed effects. In practice, the authors compare workers within the same company in the same month, thus isolating the AI exposure effect from any other factor affecting the company: budget cuts, sector slowdown, strategy changes.</p>
<p>The result: even controlling for all these factors, young workers in AI-exposed roles show a relative decline of <strong>16%</strong> compared to colleagues in non-exposed roles at the same company.</p>
<p>This kind of rigor is rare in corporate evaluations. When an AI project launches and costs drop, it’s easy to credit the AI. But maybe costs would have dropped anyway due to seasonal factors. Maybe the team was already optimizing before the launch. Maybe the comparison is with an anomalous period.</p>
<p>The solution is to define baselines and control groups before launch. Don’t compare “before vs after” but “treated vs untreated” in the same period. Use A/B tests where possible, or at least comparisons with teams, regions, or segments that haven’t adopted AI.</p>
<hr />
<h2>Toward high-frequency economic dashboards</h2>
<p>In his predictions for 2026, Brynjolfsson proposed the idea of <em>“AI economic dashboards”</em>, tools that track AI’s economic impact in near real-time, updated monthly instead of with the typical delays of official statistics.</p>
<p>It’s an ambitious proposal at the macro level. But the underlying logic is applicable at the company level: stop waiting for quarterly reports to understand if AI is working and instead build continuous monitoring systems that capture effects as they emerge.</p>
<p>Most AI projects are evaluated like traditional investments: ex-ante business case, periodic reviews, final post-mortem. But AI doesn’t behave like a traditional asset. Its effects are distributed, emergent, often unexpected. A continuous monitoring system can catch drift before it becomes a problem.</p>
<p>In practice, this means working with real-time data instead of retrospective data. If the payroll system can tell you today how many people were hired yesterday in each role, you can track AI’s effect on headcount with a lag of days, not months. The same applies to tickets handled, sales closed, errors detected.</p>
<p>Another key principle: favor leading metrics over lagging ones. The actual utilization rate (how many employees actually use the AI tool every day) is a leading indicator. If it drops, there are problems before they show up in productivity numbers.</p>
<p>As the Stanford paper segments by age, corporate dashboards should segment by role, tenure, and prior performance. AI might help top performers while harming others, or vice versa.</p>
<p>Internal comparisons are also essential: teams that adopted AI vs teams that didn’t, periods with the feature active vs periods with it deactivated. These comparisons are more informative than pure time trends.</p>
<hr />
<h2>The cost of not measuring</h2>
<p>There’s a direct economic argument for investing in better measurement. The 42% of companies that abandoned AI projects in 2025 spent budget, time, and management attention only to get nothing. With better metrics, some of those projects would have been stopped earlier. Others would have been corrected mid-course. Others still would never have started.</p>
<p>The MIT NANDA report estimates that companies are spending <strong>$30-40 billion per year</strong> on generative AI. If 95% generates no measurable ROI, we’re talking about tens of billions burned. Not because the technology doesn’t work, but because it’s applied poorly, measured worse, and therefore never corrected.</p>
<p>The Brynjolfsson paper offers a model of what AI measurement could be. Administrative data instead of surveys. Demographic granularity instead of aggregate averages. Rigorous controls instead of naive comparisons. Continuous monitoring instead of point-in-time evaluations.</p>
<p>No company has Stanford’s resources or access to ADP’s data. But the principles are transferable: segment, distinguish substitutive from augmentative use, control for confounding factors, monitor in real time. Those who adopt these principles will have an information advantage over those who continue tracking deployments and hours saved.</p>
<hr />
<h2>Sources</h2>
<p>Brynjolfsson, E., Chandar, B., &amp; Chen, R. (2025). <a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/"><em>Canaries in the Coal Mine: Six Facts about the Recent Employment Effects of AI</em></a>. Stanford Digital Economy Lab.</p>
<p>Deloitte AI Institute. (2025). <a href="https://www2.deloitte.com/us/en/pages/consulting/articles/state-of-generative-ai-in-enterprise.html"><em>State of Generative AI in the Enterprise</em></a>. Deloitte.</p>
<p>MIT Project NANDA. (2025). <a href="https://projectnanda.org/"><em>The GenAI Divide 2025</em></a>. Massachusetts Institute of Technology.</p>
<p>MIT Sloan Management Review. (2024). <a href="https://sloanreview.mit.edu/projects/the-future-of-strategic-measurement-enhancing-kpis-with-ai/"><em>The Future of Strategic Measurement: Enhancing KPIs With AI</em></a>. MIT Sloan.</p>
<p>S&amp;P Global Market Intelligence. (2025, October). <a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/2025/10/generative-ai-shows-rapid-growth-but-yields-mixed-results"><em>Generative AI Shows Rapid Growth but Yields Mixed Results</em></a>. S&amp;P Global.</p>
]]></content:encoded>
      <category>Business</category>
      <category>Research</category>
      <category>Methodology</category>
      <category>KPI</category>
      <category>Metrics</category>
      <category>AI Measurement</category>
      <category>Enterprise AI</category>
      <category>ROI</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/business/misurare-ia/"/>
    </item>
    <item>
      <title>AI Sovereignty: Europe&apos;s Decision Point</title>
      <link>https://ireneburresi.dev/en/blog/governance/ia-sovranit%C3%A0/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/governance/ia-sovranit%C3%A0/</guid>
      <pubDate>Sat, 20 Dec 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Sovereign clouds, national models, and US hyperscalers define three ideas of AI sovereignty. Europe must decide which infrastructure and governance to fund.</p>]]></description>
      <content:encoded><![CDATA[<h2>Full English Translation Coming Soon</h2>
<p>This comprehensive analysis of AI sovereignty, European choices, and geopolitical implications will be fully translated soon.</p>
<p>The article covers:</p>
<ul>
<li>Four definitions of AI sovereignty (legality, economic competitiveness, national security, value alignment)</li>
<li>Two operational models for achieving sovereignty</li>
<li>Gulf states’ AI infrastructure investments</li>
<li>Europe’s position between American providers and sovereign models</li>
<li>GAIA-X and EU cloud sovereignty initiatives</li>
</ul>
<p>For now, please refer to the Italian version for the complete content.</p>
]]></content:encoded>
      <category>Governance</category>
      <category>Business</category>
      <category>Other</category>
      <category>AI Sovereignty</category>
      <category>EU AI Act</category>
      <category>GAIA-X</category>
      <category>Geopolitics</category>
      <category>Europe</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/governance/ia-sovranit%C3%A0/"/>
    </item>
    <item>
      <title>Why replacing people doesn&apos;t fix the team</title>
      <link>https://ireneburresi.dev/en/blog/methodology/debugging-organizzativo-hackman/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/methodology/debugging-organizzativo-hackman/</guid>
      <pubDate>Sat, 22 Mar 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Not "who isn't working" but "what in the structure isn't working." Six misdiagnoses flipped using Hackman's framework. The first debug is on the structure.</p>]]></description>
      <content:encoded><![CDATA[<p>Marco’s team isn’t working. Everyone knows it. The diagnosis comes fast: “Marco isn’t motivated.” “Sara doesn’t communicate.” The plan: a one-on-one with Marco to figure out what’s blocking him, a team building workshop, and if things don’t improve, swap someone out.</p>
<p>Three months later, Marco is gone. Luca replaced him. The team still isn’t working.</p>
<p>The scene keeps repeating because the diagnosis is wrong. Not wrong in an obvious way — wrong in a plausible way, which is worse. “Marco isn’t motivated” sounds reasonable. Everyone nods. Too bad the problem isn’t Marco.</p>
<p>Social psychology has a name for this: <em>fundamental attribution error</em>. Ross described it in 1977: when we observe behavior, we tend to attribute it to personal traits (laziness, poor collaboration) while underestimating the weight of the situation the person is in.</p>
<p>J. Richard Hackman spent forty years studying teams: flight crews, orchestras, intelligence squads. The bottom line of his research: structural conditions explain up to 80% of variance in team effectiveness (<a href="https://doi.org/10.1177/0021886305281984">Wageman, Hackman &amp; Lehman, 2005</a>). When a team isn’t working, the most likely problem isn’t who’s in it. It’s how it was designed.</p>
<p>Before replacing people, check the conditions.</p>
<hr />
<h2>The five conditions: a checklist, not a theory</h2>
<p>Hackman didn’t produce an abstract model. He produced a checklist. Five conditions that, when present, make it likely a team will work. When absent, make it likely it won’t. He validated them empirically across hundreds of teams in different contexts. They work as a diagnostic tool long before they work as theory.</p>
<p>Here’s a quick translation into software context. For the full picture, start with the <a href="https://ireneburresi.dev/blog/methodology/hackman-real-team-software/">previous piece on Hackman’s team vs. working group distinction</a>.</p>
<p><strong>1. Being a real team.</strong> Clear boundaries, stable composition, real interdependence. If even one of these three is missing, you don’t have a team. You have a group of people with a shared manager.</p>
<p><strong>2. A compelling direction.</strong> A clear, challenging, meaningful objective. Not “close the sprint stories” — that’s a unit of measurement, not a direction. A compelling direction answers the question: why does this work matter?</p>
<p><strong>3. An enabling structure.</strong> Roles, norms, skills. The minimum infrastructure so people don’t have to renegotiate everything from scratch every week.</p>
<p><strong>4. A supportive organizational context.</strong> Does the team have the information it needs? The resources? Does the reward system incentivize group outcomes or individual performance? This is the condition the team can’t give itself. It depends on the organization around it.</p>
<p><strong>5. Competent process coaching.</strong> Not technical mentoring: someone who helps the team work better <em>together</em>. How they make decisions, handle disagreements, distribute work. It’s the condition that matters least of the five (the 10% in Hackman’s 60/30/10), but it’s the one everyone focuses on.</p>
<p>The instinct when managing teams is to start at the bottom: coaching, facilitation, interpersonal dynamics. Hackman says: start at the top.</p>
<hr />
<h2>The table that flips the diagnosis</h2>
<p>This is the core of the argument. Six phrases I hear repeated constantly. For each one: the usual diagnosis, the most likely structural cause, and where to look instead.</p>
<p>Some of these mappings are directly anchored to Hackman’s research. Others are my translation into the software context, and I flag them as such.</p>
<h3>“Marco isn’t motivated”</h3>
<p>It’s Marco’s problem, people say. He lacks drive, ownership. Maybe he’s not right for the role.</p>
<p>More likely, the team lacks a <strong>compelling direction</strong> (condition 2). If the team’s objective is vague, purely administrative, or changes every two weeks, motivation isn’t a trait of Marco’s. It’s a rational response to a context that gives no reason to invest energy. I’ve seen brilliant developers seem disengaged because their team’s mandate was “support the product” — which in practice meant answering tickets endlessly with no vision of where things were heading.</p>
<p>Don’t look at Marco. Look at the team’s mandate. If you can’t explain in one sentence why the work matters, that’s your problem.</p>
<h3>“Sara doesn’t communicate”</h3>
<p>Where to look here is at work design. If everyone works on a separate feature and interactions are limited to the occasional comment on a pull request, there’s no structural reason to communicate more. Sara isn’t uncollaborative. She’s rational: why would she update others on work that doesn’t affect them?</p>
<p>The cause is a lack of <strong>real interdependence</strong> (condition 1). You can run all the communication workshops you want: if the work structure doesn’t create mutual dependencies, communication will remain an empty ritual.</p>
<h3>“The team doesn’t trust each other”</h3>
<p>The classic response: team building activities, vulnerability exercises, an offsite to get to know each other better.</p>
<p>Trust in a team doesn’t come from an afternoon workshop. It comes from working together long enough to predict how the other person will behave. Hackman saw this with flight crews: crews that had flown together for a while made fewer errors than newly formed ones, even with less experienced pilots. The most likely problem is <strong>unstable composition</strong> (condition 1). If you rotate people between teams every quarter, you restart from zero each time.</p>
<p>How many times has the team’s composition changed in the past year? If the answer is “often,” the trust deficit isn’t a relationship problem. It’s structural turnover.</p>
<h3>“Retros don’t produce anything”</h3>
<p>Here the question comes before the answer: is there a shared process to improve? Retrospectives assume a team that collaborates on a common outcome and wants to improve how they do it. If everyone has their own workflow, priorities, and blockers, the retro has no object. The format is irrelevant. The cause is that the group isn’t a <strong>real team</strong> (condition 1).</p>
<h3>“The lead can’t guide the team”</h3>
<p>The usual diagnosis: they lack soft skills. Send them to a leadership course.</p>
<p>The lead might be excellent. But if the team doesn’t have access to the information it needs, if resources get cut without notice, if the evaluation system rewards individual performance and ignores group outcomes, the lead is in an impossible position. It’s like asking a pilot to land well with a poorly designed airplane — you can send them to every flight course there is, but if the flaps don’t work, the problem isn’t their technique.</p>
<p>The cause is the <strong>organizational context</strong> (condition 4). Does the lead have the authority to make decisions? Are priorities stable enough to allow a plan? If not, the leverage is in the organization, not the person.</p>
<h3>“Too much conflict”</h3>
<p>Here the answer is less straightforward. Conflict in a team can be a good sign: if the team is real and negotiating norms for working together (condition 3), some friction is physiological. Hackman says it clearly: the best teams aren’t conflict-free. They’re teams that have learned to manage conflict.</p>
<p>But conflict can also signal <strong>unclear boundaries</strong> (condition 1): who decides what? Who has authority over which part of the system? If two people both think they’re responsible for the same area, the conflict isn’t relational. It’s structural.</p>
<p>The useful distinction: if the conflict is about <em>how</em> to do things, it’s probably healthy. If it’s about <em>who</em> should do what, it’s a boundary problem. And if it’s chronic despite everyone’s best intentions, it’s almost certainly not a people problem.</p>
<hr />
<h2>Why we keep getting the diagnosis wrong</h2>
<p>If structural conditions matter this much, why does the default remain “someone’s fault”?</p>
<p>The first reason is that <strong>structure is invisible</strong>. You see people. You see behaviors. You don’t see the conditions in which those behaviors emerge. Hackman pressed this point in the last paper of his career, “From causes to conditions” (<a href="https://onlinelibrary.wiley.com/doi/full/10.1002/job.1774">Hackman, 2012</a>): research on teams has focused too much on internal causes (who does what, how they behave) and too little on external conditions (how the team is designed, what context it operates in). The same applies to people managing teams in practice.</p>
<p>The second is that <strong>replacing people feels easier</strong>. Giving Marco feedback, sending Sara to a workshop, swapping the lead: these are all actions a manager can take tomorrow morning. Redesigning the team’s mandate, stabilizing composition, changing the incentive system — those require time, authority, and often negotiation with someone above. The interpersonal diagnosis is attractive because it has solutions at hand. Too bad they’re the wrong solutions.</p>
<p>The third is more insidious, and I’ve seen it up close: <strong>the organization incentivizes individual diagnosis</strong>. Performance reviews evaluate people, not conditions. PIPs apply to people. “Cultural fit” is an attribute of people. The entire management apparatus is built around the idea that performance is an individual trait. Admitting the problem is structural means admitting the system is poorly designed — and the person who designed the system is often the one making the diagnosis.</p>
<hr />
<p>Next time someone says “the problem is Marco,” pause for a moment. Not because Marco is necessarily innocent: sometimes the problem really is the person. But it’s less common than we think, and the interpersonal diagnosis is so intuitive we always get there first.</p>
<p>Try reframing. Not “who isn’t working?” but “what in the structure isn’t working?” Move from person to condition. Then ask yourself: is this condition under my control? If yes, that’s where the energy goes. If not, at least you know that no team building workshop will fix it.</p>
<p>It’s not a small difference. It’s the difference between debugging the code and debugging the compiler.</p>
<hr />
<h2>Sources</h2>
<ul>
<li>Hackman, J.R. (2002). <em>Leading Teams: Setting the Stage for Great Performances</em>. Harvard Business School Press.</li>
<li>Hackman, J.R. (2011). <em>Collaborative Intelligence: Using Teams to Solve Hard Problems</em>. Berrett-Koehler.</li>
<li>Hackman, J.R. (2012). <a href="https://onlinelibrary.wiley.com/doi/full/10.1002/job.1774">From causes to conditions in group research</a>. <em>Journal of Organizational Behavior</em>, 33, 428-444.</li>
<li>Wageman, R., Hackman, J.R. &amp; Lehman, E.V. (2005). <a href="https://doi.org/10.1177/0021886305281984">Team Diagnostic Survey: Development of an Instrument</a>. <em>Journal of Applied Behavioral Science</em>, 41(4), 373-398.</li>
<li>Ross, L. (1977). The intuitive psychologist and his shortcomings: Distortions in the attribution process. In L. Berkowitz (Ed.), <em>Advances in Experimental Social Psychology</em> (Vol. 10, pp. 173-220). Academic Press.</li>
</ul>
]]></content:encoded>
      <category>Methodology</category>
      <category>Business</category>
      <category>Team Design</category>
      <category>Hackman</category>
      <category>Organizational Design</category>
      <category>Team Effectiveness</category>
      <category>Fundamental Attribution Error</category>
      <category>Diagnostics</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/methodology/debugging-organizzativo-hackman/"/>
    </item>
    <item>
      <title>Frustrated with Agile? Maybe your team isn&apos;t actually a team</title>
      <link>https://ireneburresi.dev/en/blog/methodology/hackman-real-team-software/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/methodology/hackman-real-team-software/</guid>
      <pubDate>Sat, 15 Mar 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Hackman's distinction between real teams and working groups explains why standups, retros, and planning feel pointless. The problem isn't Agile — it's a structural mismatch.</p>]]></description>
      <content:encoded><![CDATA[<p>The standup takes twelve minutes. Five people, each reciting their update while staring at some vague point on the screen. Nobody comments on what anyone else says, because there’s no reason to: everyone is working on their own feature, in a different corner of the codebase. The standup ends, everyone goes back to doing exactly what they would have done without it. Then the retro. Three sticky notes come out of it: “improve communication,” same as last month. And sprint planning, which is really just individual task assignment with a two-week timer.</p>
<p>If this sounds familiar, you’re not alone. The <a href="https://digital.ai/resource/state-of-agile-report/">17th State of Agile Report</a> by <a href="https://digital.ai/">Digital.ai</a> (2024) found that only 11% of practitioners report being “very satisfied” with Agile practices in their organization. The frustration runs deep enough that two original signatories of the Agile Manifesto have publicly turned against what Agile has become: Ron Jeffries, co-creator of Extreme Programming, wrote that developers should <a href="https://ronjeffries.com/articles/018-01ff/abandon-1/">abandon Agile</a>, or at least the version organizations have made of it. Dave Thomas, another signatory, declared that <a href="https://pragdave.me/thoughts/active/2014-03-04-time-to-kill-agile.html">Agile is dead</a>, hollowed out by marketing and mass certification.</p>
<p>The most common diagnosis is that Agile has become bureaucracy: too many rituals, too much process, not enough code. It makes sense. Who hasn’t thought that at least once, walking out of yet another endless planning session?</p>
<p>But there’s another possibility. One that has nothing to do with Agile itself, but with something more fundamental: the structure you’re applying it to.</p>
<p>J. Richard Hackman spent 40 years studying real teams: flight crews, orchestras, intelligence teams, surgical teams. His research, condensed in <em>Leading Teams</em> (2002) and <em>Collaborative Intelligence</em> (2011), arrives at a distinction that almost nobody in software makes: the distinction between a <strong>team</strong> and a <strong>working group</strong>. They are different things. They work differently. And they require different tools.</p>
<p>Agile practices are designed for teams. Apply them to a working group and you get exactly the frustration you’re feeling. The problem isn’t the method. It’s a structural mismatch.</p>
<hr />
<h2>The most overused word in software</h2>
<p>In software, “team” is the default word for any group of people working on the same project. Five developers sharing a Jira board? Team. Two backend engineers, a frontend dev, and a designer reporting to the same manager? Team. Eight people across three time zones who meet at the 9 AM standup? Team.</p>
<p>Hackman would disagree. In <em>Leading Teams</em> he defines a “real team” through three minimum properties, all required.</p>
<p>The first is <strong>clear boundaries</strong>: everyone knows who is on the team and who is not. Sounds obvious, but in software practice it isn’t. The designer “shared” across three teams — in or out? The developer “on loan” for two sprints — a member? If you can’t make the list without hesitating, the boundaries aren’t clear.</p>
<p>The second is <strong>stable composition</strong>. People stay the same long enough to develop shared ways of working. Hackman studied flight crews: NASA data, which he reports in <em>Leading Teams</em>, showed that newly formed crews made more errors than those who had been flying together for a while, even when the newer crews had more experienced pilots. In software, quarterly rotation of people between teams destroys this effect. Every time, you start from zero.</p>
<p>The third, and most underrated, is <strong>real interdependence</strong>. The team’s output depends on collaboration between members — it’s not the sum of individual contributions. This is where most software “teams” fall apart. Five developers working on five independent features, with five cross-reviews done as a formality, are not interdependent. One person’s work doesn’t change another’s. If you removed all the meetings and put them in separate rooms, the result would be the same.</p>
<p>The test is brutal in its simplicity: if you eliminated standups, retros, and planning tomorrow, and everyone worked on their own, would the final product suffer? If the answer is no — if the result would be identical, maybe even faster without the interruptions — then what you have is not a team. It’s a group of individuals with a shared manager.</p>
<p>That’s not an insult. It’s a diagnosis. And the diagnosis is the first step toward stopping the use of the wrong tools.</p>
<hr />
<h2>Structural conditions: the 60/30/10</h2>
<p>Hackman didn’t stop at the definition. He studied what makes a team effective, and the answer is less intuitive than you’d think.</p>
<p>Research by Hackman and his collaborator Ruth Wageman identifies five conditions that enable team performance: being a real team, having a compelling direction, an enabling structure, a supportive organizational context, and competent coaching. I won’t go deep on all of them here — each deserves its own article. The relevant point is different: how much do these conditions matter compared to what a leader does day to day?</p>
<p>Hackman’s answer, laid out in <em>Collaborative Intelligence</em> (2011), is the <strong>60/30/10</strong>: 60% of a team’s effectiveness depends on design — the structural conditions put in place before the team starts working. 30% depends on launch — how the team is kicked off, the first days, the initial norms. The remaining 10% on ongoing coaching.</p>
<p>A clarification: Hackman presents this split as a “best estimate,” not as the result of a single study yielding those exact percentages. It’s a heuristic that synthesizes decades of research. Not a precise data point.</p>
<p>But the order of magnitude has solid empirical grounding. The Team Diagnostic Survey (TDS), developed by Wageman, Hackman, and Lehman and published in 2005, was administered to 2,474 people across 321 teams. The finding: structural conditions explain up to 80% of the variance in team effectiveness (<a href="https://doi.org/10.1177/0021886305281984">Wageman, Hackman &amp; Lehman, 2005</a>). Not 20%. Not 50%. Eighty percent.</p>
<p>An earlier study by Wageman (2001) on <a href="https://doi.org/10.1287/orsc.12.5.559.10094">43 self-managing teams at Xerox</a> had already shown the same pattern: a leader’s design activities influenced team performance. Day-to-day coaching activities did not.</p>
<p>The implication for software is direct. When a team isn’t working, the instinctive reaction is to work on dynamics: facilitate retros better, improve communication, do team building. Hackman’s framework suggests the most powerful lever is upstream. Who is on the team? What is the mandate? How is the work designed? Does the organizational context support it? If these conditions aren’t there, no amount of facilitation will compensate.</p>
<p>But the first of those five conditions — “being a real team” — opens a question that almost nobody in software asks.</p>
<hr />
<h2>Team or working group: the distinction that changes everything</h2>
<p>A point Hackman makes often, and that gets misunderstood just as often: the distinction between team and working group is not a value judgment. It’s not “team = good, working group = bad.” They are two different organizational modes, each with its own strengths.</p>
<p>A <strong>working group</strong> is a set of people reporting to the same manager who may coordinate, but whose output is primarily individual. Everyone has their own goals, responsibilities, deliverables. The manager coordinates, assigns, removes obstacles. Deep interdependence is not required.</p>
<p>A <strong>team</strong> produces collective output. The result cannot be decomposed into the sum of individual contributions: it requires continuous collaboration, shared decisions, mutual adjustment. The cost is higher — you need stable boundaries, shared norms, time to develop ways of working together. But for certain kinds of problems, it’s the only configuration that works.</p>
<p>The damage comes from confusing the two.</p>
<p>A working group managed as a team generates overhead without benefit. Coordination meetings exist, but there’s nothing substantial to coordinate. Retrospectives don’t produce actions because there’s no shared process to improve: everyone has their own workflow, priorities, blockers. The Agile ceremony becomes a fixed cost on work that doesn’t require it.</p>
<p>The damage goes both ways, though. A real team managed as a working group is equally dysfunctional. If you have people who need to collaborate on a complex problem and treat them as individual contributors — assigning separate tasks, evaluating them individually, without protecting time for joint work — you’re undermining the one thing that makes that team effective: interdependence.</p>
<p>I’ve seen both scenarios. The first is more common in software, because the default organization tends to be the working group (developers assigned to individual features), while the process infrastructure is almost always that of a team (Scrum, Kanban with standups, retros, planning).</p>
<p>The operational question isn’t “how do I improve my team?” It’s more fundamental: is what I’m leading a team or a working group? The answer changes everything that follows.</p>
<hr />
<h2>The Agile mismatch: right tools, wrong structure</h2>
<p>Back to the frustration we started with. Agile practices — Scrum in particular — didn’t emerge in a vacuum. They’re designed around a specific assumption: that a small group of people works interdependently toward a shared goal, iterating together. The sprint assumes a common goal. The standup assumes that knowing what others are doing changes your work. The retro assumes a shared process to inspect. Planning assumes collective prioritization decisions.</p>
<p>These are all tools that assume interdependence. Without it, they lose their point.</p>
<p>Now look at the typical “team” setup in many software organizations. People are assigned — often rotated — to a “team” that is really an organizational container. Everyone works on their own user story, with interactions limited to code review and the occasional Slack question. The “sprint goal” is the sum of individual stories. Interdependence is minimal or absent.</p>
<p>Apply team tools to this structure and the result is predictable. The standup becomes a round of updates nobody listens to. The retro produces generic complaints. Planning becomes bureaucracy. Not because Scrum is bureaucracy — but because you’re using it on a structure it wasn’t designed for.</p>
<p>The Manifesto signatories say as much, in different words. Jeffries talks about “Dark Scrum” — organizations using Agile rituals as control mechanisms, draining them of collaboration. Thomas says “Agile” was turned into a commercial noun when it was meant to be an adjective describing a way of working. Their critique is legitimate. But the diagnosis they offer — “organizations have corrupted Agile” — is incomplete. Hackman’s framework provides a more structural one: many organizations haven’t corrupted Agile. They’ve applied it to structures that aren’t teams.</p>
<p>I bring this up because I’ve seen it happen more times than I’d like to admit, including in contexts with competent people and good intentions. The standard response to the malaise was always the same: change the facilitator, try a different retro format, add a ceremony. Hackman’s framework gave me the vocabulary for what I felt but couldn’t articulate: the problem wasn’t the execution. It was the premise.</p>
<p>This changes the diagnosis and the solutions. If the problem is “Agile has become bureaucracy,” the answer is less process, fewer rituals, more autonomy. Sometimes that’s right. But if the problem is a structural mismatch, the answer is different: either transform the structure into a real team — with the costs that entails — or accept that you have a working group and adopt tools consistent with that reality. Both options are legitimate. What doesn’t work is staying in the middle.</p>
<hr />
<p>Back to the twelve-minute standup we started with. Five people, five updates, no interaction. The obvious diagnosis is that the standup is poorly facilitated, or that Scrum doesn’t work, or that the team needs to “work on communication.” These are all responses that address the symptom.</p>
<p>Hackman’s question is different, and it comes first: do those five people need to talk to each other every morning to do their work? If each person’s work doesn’t depend on the others’, the standup isn’t poorly facilitated. It’s useless by design. The retro doesn’t produce actions because there’s no shared process to improve. Planning is bureaucracy because there are no collective decisions to make.</p>
<p>It’s not a question of execution. It’s a question of structure.</p>
<p>Next time you think “Agile doesn’t work,” try reframing. Don’t ask how to improve the process. Ask whether what you’re leading is a team or a working group. The answer isn’t a judgment. It’s a diagnosis. And from the diagnosis, everything else follows.</p>
<hr />
<h2>Sources</h2>
<ul>
<li>Hackman, J.R. (2002). <em>Leading Teams: Setting the Stage for Great Performances</em>. Harvard Business School Press.</li>
<li>Hackman, J.R. (2011). <em>Collaborative Intelligence: Using Teams to Solve Hard Problems</em>. Berrett-Koehler.</li>
<li>Wageman, R. (2001). <a href="https://doi.org/10.1287/orsc.12.5.559.10094">How Leaders Foster Self-Managing Team Effectiveness: Design Choices Versus Hands-on Coaching</a>. <em>Organization Science</em>, 12(5), 559-577.</li>
<li>Wageman, R., Hackman, J.R. &amp; Lehman, E.V. (2005). <a href="https://doi.org/10.1177/0021886305281984">Team Diagnostic Survey: Development of an Instrument</a>. <em>Journal of Applied Behavioral Science</em>, 41(4), 373-398.</li>
<li><a href="https://digital.ai/">Digital.ai</a> (2024). <a href="https://digital.ai/resource/state-of-agile-report/">17th State of Agile Report</a>.</li>
<li>Jeffries, R. (2018). <a href="https://ronjeffries.com/articles/018-01ff/abandon-1/">Developers Should Abandon Agile</a>.</li>
<li>Thomas, D. (2014). <a href="https://pragdave.me/thoughts/active/2014-03-04-time-to-kill-agile.html">Agile is Dead (Long Live Agility)</a>.</li>
</ul>
]]></content:encoded>
      <category>Methodology</category>
      <category>Business</category>
      <category>Team Design</category>
      <category>Agile</category>
      <category>Hackman</category>
      <category>Team vs Working Group</category>
      <category>Scrum</category>
      <category>60-30-10</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/methodology/hackman-real-team-software/"/>
    </item>
    <item>
      <title>Who Will the Senior Engineers of Tomorrow Be?</title>
      <link>https://ireneburresi.dev/en/blog/business/senior-domani/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/business/senior-domani/</guid>
      <pubDate>Mon, 06 Jan 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>Employment for developers under 25 dropped 20% since ChatGPT's launch. Companies hire fewer juniors because AI does those tasks. But without junior developers today, who will lead teams in ten years?</p>]]></description>
      <content:encoded><![CDATA[<h2>The apprenticeship paradox</h2>
<p><em>Companies don’t hire juniors because AI does those tasks better. But without junior developers today, who will lead teams in ten years?</em></p>
<p>There’s a question that rarely appears in quarterly reports: if we stop hiring people who are learning, who will know how to do this job a decade from now?</p>
<p>The numbers tell a story that should concern anyone managing technical teams. Employment for software developers between ages 22 and 25 has dropped 20% from the peak in late 2022, according to a <a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/">paper from Stanford’s Digital Economy Lab</a> based on payroll data from 25 million workers. It’s not a uniform decline: in the same period, employment for those over 35 in the same roles grew 8%.</p>
<p>The mechanism is what we might call the apprenticeship paradox: companies stop hiring entry-level because AI does those tasks better than a recent graduate. But without entry-level workers today, they won’t have senior engineers tomorrow.</p>
<hr />
<h2>The numbers of collapse</h2>
<p>The contraction is not an impression. It’s documented by multiple independent sources.</p>
<p>Entry-level hiring at the top 15 tech companies dropped 25% between 2023 and 2024, according to <a href="https://spectrum.ieee.org/ai-effect-entry-level-jobs">SignalFire</a>. Since 2021, the average age of technical hires has increased by three years. Companies aren’t just hiring less: they’re hiring differently, preferring senior profiles who can be productive from day one.</p>
<p>Tech internships have collapsed 30% since 2023, <a href="https://stackoverflow.blog/2025/12/26/ai-vs-gen-z">according to Handshake</a>. Meanwhile, applications have increased 7%. More people competing for fewer positions, and the remaining positions require increasingly prior experience.</p>
<p>A <a href="https://www.finalroundai.com/blog/ai-is-making-it-harder-for-junior-developers-to-get-hired">Harvard study</a> of 285,000 American companies found that when firms adopt generative AI, junior employment drops 9-10% within six quarters. Senior employment remains stable. These aren’t mass layoffs: it’s a silent hiring freeze. Companies simply stop opening entry-level positions.</p>
<p>The pattern repeats in Europe. Junior tech positions have <a href="https://restofworld.org/2025/engineering-graduates-ai-job-losses/">dropped 35%</a> in major EU countries during 2024, based on aggregated data from LinkedIn, Indeed, and Eures. In the UK, the Big Four consulting firms cut graduate hiring between 6% and 29% in two years. In India, IT companies have reduced entry-level roles by 20-25%, according to an EY report.</p>
<p>The World Economic Forum, in its Future of Jobs Report 2025, warns that 40% of employers expect to reduce staffing where AI can automate tasks. And automatable tasks are, almost by definition, the ones junior developers used to do.</p>
<hr />
<h2>The logic of the short term</h2>
<p>The rationale behind these choices is understandable. A senior engineer with AI tools can do what previously required two or three juniors, at least for certain tasks. GitHub Copilot, Cursor, and similar tools promise productivity gains of 20-50% according to their vendors. For a CFO looking at the next quarter, hiring a junior who will need six months of training before being productive seems like a difficult investment to justify.</p>
<p>James O’Brien, a computer science professor at Berkeley who works with startups, <a href="https://sfstandard.com/2025/05/20/silicon-valley-white-collar-recession-entry-level/">describes the shift</a>: “Previously, startups would hire one senior person and two or three early-career coders to assist. Now they ask: why hire a recent graduate when AI is cheaper and faster?”</p>
<p>It’s a reasonable question in the short term. Code generated by AI isn’t top quality, but neither is code written by a recent graduate. The difference, O’Brien notes, is that the iterative process to improve AI code takes minutes. A junior might take days for the same task.</p>
<p>Heather Doshay, head of talent at SignalFire, sums it up: “Nobody has the patience or time for hand-holding in this new environment, where much of the work can be done autonomously by AI.”</p>
<hr />
<h2>The problem nobody calculates</h2>
<p>There’s a flaw in this logic, and it’s called the talent pipeline.</p>
<p>Matt Garman, CEO of AWS, <a href="https://www.finalroundai.com/blog/aws-ceo-ai-cannot-replace-junior-developers">said it explicitly</a>: “If you don’t have a talent pipeline you’re building, if you don’t have junior people you’re mentoring and growing in the company, we often find that’s where the best ideas come from. If a company stops hiring juniors and developing them, eventually the whole system falls apart.”</p>
<p>It’s not rhetoric. It’s demographic mathematics applied to organizations. Every senior engineer, every tech lead, every CTO was once a junior. The path from recent graduate to technical leader requires years of experience on real projects, mistakes made and corrected, feedback received, patterns internalized. There is no shortcut.</p>
<p>If the industry stops hiring juniors in 2023, by 2033 it will have a structural shortage of mid-level talent. By 2038, there will be a shortage of senior engineers. By 2043, there will be no one to promote to technical leadership roles.</p>
<p>The problem is that this cost doesn’t appear in any quarterly balance sheet. It’s an invisible debt that accumulates silently, and when it becomes obvious, it will be too late to remedy quickly.</p>
<hr />
<h2>AI that teaches and AI that atrophies</h2>
<p>There’s a further irony in this situation. The same AI tools that are eliminating junior roles could, in theory, accelerate learning. An AI tutor available 24/7, patient, answering every question: it sounds like every student’s dream.</p>
<p>The reality is more complicated.</p>
<p>An experiment conducted by Wharton and Penn researchers on nearly a thousand high school math students tested two versions of a GPT-4-based tutor. The group with access to a ChatGPT-like interface (GPT Base) achieved 48% better results during assisted practice sessions. The group with a tutor designed to guide without giving direct answers (GPT Tutor) achieved 127% better results.</p>
<p>But here’s the point: when AI was removed and students took the exam on their own, the GPT Base group achieved 17% worse results than the control group who never used AI. The GPT Tutor group, by contrast, achieved results similar to control.</p>
<p>Students were using AI as a crutch. They performed better with assistance but learned less. When the assistance was removed, they found themselves worse off than those who never had it.</p>
<p>A <a href="https://time.com/7295195/ai-chatgpt-google-learning-school/">study from MIT Media Lab</a> documented what researchers call “cognitive debt”: using LLMs for writing seems to reduce mental effort during the task, but at the cost of more superficial learning. Researcher Nataliya Kosmyna expressed concern about developing brains: “Developing brains are the ones at highest risk.”</p>
<p>It doesn’t mean AI can’t help learning. The Wharton study shows it can, if designed with the right safeguards. But “wild” AI, the kind that gives answers instead of guiding toward answers, can do damage.</p>
<hr />
<h2>The new profile of the junior</h2>
<p>If fewer juniors will be hired, what characteristics must they have to be hired?</p>
<p>Market signals are clear. It’s no longer enough to know how to code. Employers <a href="https://spectrum.ieee.org/ai-effect-entry-level-jobs">expect</a> recent graduates to be able to manage projects, communicate with clients, understand the software development lifecycle. The “grunt work” that once served as a training ground is being automated. Those entering must be operational at a higher level almost from day one.</p>
<p>Jamie Grant, who manages career services for engineering at the University of Pennsylvania, describes the change: “They’re not necessarily just programming. There’s much more high-level thinking and understanding of the software development lifecycle.”</p>
<p>David Malan of Harvard, who teaches the world’s most-followed introduction to programming course, notes that the biggest impact of AI has been on programmers, not on roles that were expected (like call centers). The reason: programming work is relatively solitary and highly structured, perfect for automation.</p>
<p>But Malan also notes something interesting: in the United States, employment for “programmers” dropped 27.5% between 2023 and 2025, but employment for “software developers,” a more design-oriented position, dropped only 0.3%. The difference is in the level of abstraction. Those who write code are vulnerable. Those who design systems less so.</p>
<hr />
<h2>Three scenarios for the future</h2>
<p><strong>Scenario 1: The collapse of the pipeline</strong></p>
<p>Companies continue not to hire juniors. In five to ten years, the shortage of mid-level talent becomes acute. The remaining senior engineers command astronomical salaries. Companies that can’t afford them lose competitiveness. The industry polarizes between a few giants who can attract talent and everyone else struggling.</p>
<p><strong>Scenario 2: Apprenticeship reinvented</strong></p>
<p>Some companies realize the problem is coming and invest against the trend. They create intensive training programs, perhaps assisted by AI designed to teach instead of replace. They become the preferred employers for top talent, who know they can grow there. In the long term, they have a competitive advantage.</p>
<p><strong>Scenario 3: Uneven democratization</strong></p>
<p>AI lowers the barrier to entry for some skills (writing working code) but raises it for others (designing systems, debugging complex problems, managing AI itself). Those with access to quality training and mentorship can skip some steps. Those without remain stuck. Inequality of opportunity increases.</p>
<p>None of these scenarios is inevitable. They are possibilities that depend on choices companies, educational institutions, and policymakers will make in the coming years.</p>
<hr />
<h2>What those who hire can do</h2>
<p>If you manage a team or influence hiring decisions, some questions deserve reflection.</p>
<p><strong>Are you optimizing for the next quarter or the next ten years?</strong> A junior costs more in the short term. But the alternative is to depend entirely on the external market for talent, competing with everyone else who made the same choice.</p>
<p><strong>Is your team still teaching?</strong> If senior people spend all their time producing and no one teaching, you’re consuming human capital without regenerating it.</p>
<p><strong>How do you use AI in training?</strong> If your juniors use Copilot to get answers instead of learning to find them, you’re accelerating their short-term productivity while compromising their long-term growth.</p>
<p><strong>Are you hiring for today’s skills or tomorrow’s adaptability?</strong> Specific technical skills have an increasingly short half-life. The ability to learn, to reason about new problems, to work with people—those last.</p>
<hr />
<h2>What those starting out can do</h2>
<p>If you’re early in your career in a market that seems to close doors on you, some principles can help.</p>
<p>AI isn’t eliminating all junior work. It’s eliminating repetitive, isolated junior work. The roles that survive require human interaction, judgment about ambiguous problems, creativity applied to specific contexts. Look for those.</p>
<p>Learn to use AI as a tool, not a crutch. The difference between using ChatGPT to get answers and using it to explore problems is the difference between atrophying and growing.</p>
<p>Networking matters more than ever. If junior positions are scarce, competition is fierce, and often the person with a connection wins, not the person with the best CV. It’s not fair, but it’s real.</p>
<p>Cross-functional skills are not optional. Communication, project management, understanding the business: these are things AI can’t do and employers seek even in technical profiles.</p>
<hr />
<h2>The unanswered question</h2>
<p>I return to the initial question: who will the senior engineers of tomorrow be?</p>
<p>I don’t have a certain answer. No one does. We’re conducting a real-time experiment, without a control group, on a global scale.</p>
<p>What I know is that every senior person I know was once a junior who someone decided to hire and train. Every tech lead made beginner mistakes that someone had the patience to correct. Every systems architect wrote embarrassing code before writing elegant code.</p>
<p>If we eliminate that phase, if we treat it as a cost to cut rather than an investment to protect, we’re not optimizing. We’re consuming capital that we don’t know how to regenerate.</p>
<p>The question isn’t whether AI can replace juniors. It can, for many tasks. The question is whether we want an industry that only knows how to consume skills or one that also knows how to produce them.</p>
<p>For now, the numbers suggest we’ve chosen the first option. The bill will come. Not next quarter. But it will come.</p>
<hr />
<h2>Sources</h2>
<p>Brynjolfsson, E., Chandar, B., &amp; Chen, R. (2025). <a href="https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/"><em>Canaries in the Coal Mine: Six Facts about the Recent Employment Effects of AI</em></a>. Stanford Digital Economy Lab.</p>
<p>Bastani, H., Bastani, O., Sungu, A., Ge, H., Kabakcı, Ö., &amp; Mariman, R. (2024). <em>Generative AI Can Harm Learning</em>. The Wharton School Research Paper.</p>
<p>Stack Overflow. (2025, December). <a href="https://stackoverflow.blog/2025/12/26/ai-vs-gen-z"><em>AI vs Gen Z: How AI has changed the career pathway for junior developers</em></a>. Stack Overflow Blog.</p>
<p>IEEE Spectrum. (2025, December). <a href="https://spectrum.ieee.org/ai-effect-entry-level-jobs"><em>AI Shifts Expectations for Entry Level Jobs</em></a>.</p>
<p>Rest of World. (2025, December). <a href="https://restofworld.org/2025/engineering-graduates-ai-job-losses/"><em>AI is wiping out entry-level tech jobs, leaving graduates stranded</em></a>.</p>
<p>Kosmyna, N., et al. (2025). <a href="https://arxiv.org/abs/2506.08872"><em>Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task</em></a>. arXiv.</p>
<p>World Economic Forum. (2025). <a href="https://www.weforum.org/publications/the-future-of-jobs-report-2025/"><em>Future of Jobs Report 2025</em></a>.</p>
<p>FinalRound AI. (2025). <a href="https://www.finalroundai.com/blog/aws-ceo-ai-cannot-replace-junior-developers"><em>AWS CEO Shares 3 Solid Reasons Why Companies Shouldn’t Replace Juniors with AI Agents</em></a>.</p>
]]></content:encoded>
      <category>Business</category>
      <category>Methodology</category>
      <category>Junior Developers</category>
      <category>AI Impact</category>
      <category>Talent Pipeline</category>
      <category>Future of Work</category>
      <category>Learning</category>
      <category>Career</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/business/senior-domani/"/>
    </item>
    <item>
      <title>From SEO to GEO: Technical Guide to AI Search Optimization</title>
      <link>https://ireneburresi.dev/en/blog/engineering/seo-vs-geo/</link>
      <guid isPermaLink="true">https://ireneburresi.dev/en/blog/engineering/seo-vs-geo/</guid>
      <pubDate>Sat, 04 Jan 2025 00:00:00 GMT</pubDate>
      <dc:creator>Irene Burresi</dc:creator>
      <dc:language>en</dc:language>
      <description><![CDATA[<p>robots.txt, llms.txt, structured data with E-E-A-T, answer blocks, multimodal: complete technical infrastructure to position yourself in generative search engines. Princeton paper, Cloudflare 2025 data, concrete implementations.</p>]]></description>
      <content:encoded><![CDATA[<h2>The paradigm shift: from links to citations</h2>
<p><em>GPTBot went from 5% to 30% of crawler traffic in one year. Traffic generated by user queries to AI has grown 15 times. Traditional SEO infrastructure no longer intercepts this flow.</em></p>
<p><strong>TL;DR:</strong> Optimization for generative search engines (GEO) requires specific technical interventions: configure robots.txt for 20+ AI crawlers, implement llms.txt to guide LLMs toward priority content, extend structured data with JSON-LD including Person schema with complete E-E-A-T (73% higher selection rate). Structure content in answer blocks of 134-167 words to facilitate extraction. Multimodal content has +156% selection rate. Princeton research shows that adding citations from authoritative sources increases visibility up to 40%. Those who implement now build competitive advantages that are difficult to catch up with.</p>
<hr />
<p>Traditional SEO optimizes for a specific goal: ranking in the sorted lists returned by search engines. The user searches, receives ten blue links, clicks. Traffic arrives.</p>
<p>Generative search engines work differently. ChatGPT, Perplexity, Gemini, Claude don’t return lists of links. They synthesize answers by drawing from multiple sources, citing (or not) the source. The user gets an answer, not a list of options.</p>
<p>According to <a href="https://blog.cloudflare.com/from-googlebot-to-gptbot-whos-crawling-your-site-in-2025/">Cloudflare data from December 2025</a>, GPTBot reached 30% of AI crawler traffic, up from 5% the previous year. Meta-ExternalAgent entered at 19%. ChatGPT-User, the bot that accesses web pages when users ask questions, registered growth of 2,825%. Traffic related to user queries increased 15 times over the course of the year.</p>
<p>This is not marginal change. It’s a new acquisition channel that requires dedicated infrastructure.</p>
<hr />
<h2>robots.txt: configuration for AI crawlers</h2>
<p>The robots.txt file communicates to crawlers which parts of the site they can access. For traditional search engines, the configuration is established. For AI crawlers, the landscape is fragmented: each provider uses different user-agents, with different purposes.</p>
<h3>Map of major AI crawlers</h3>
<p><strong>OpenAI</strong> operates with three distinct crawlers:</p>
<pre><code>User-agent: GPTBot
# Training foundational models. Collects data to train GPT.

User-agent: ChatGPT-User
# User browsing. Accesses pages when a user asks for information.

User-agent: OAI-SearchBot
# Search. Indexes content for ChatGPT's search function.
</code></pre>
<p><strong>Anthropic</strong> uses:</p>
<pre><code>User-agent: ClaudeBot
# Training and updating Claude.

User-agent: Claude-Web
# Web access for user functionality.

User-agent: anthropic-ai
# Generic Anthropic crawler.
</code></pre>
<p><strong>Perplexity</strong>:</p>
<pre><code>User-agent: PerplexityBot
# Indexing for AI answer engine.

User-agent: Perplexity-User
# Fetch for user queries.
</code></pre>
<p><strong>Google</strong> has separated functions:</p>
<pre><code>User-agent: Google-Extended
# Token for AI use. NOT a bot, it's a flag.
# Blocking this user-agent prevents use of content for AI training
# while maintaining standard indexing.

User-agent: Googlebot
# Traditional crawler for Search.
</code></pre>
<p><strong>Meta</strong>:</p>
<pre><code>User-agent: Meta-ExternalAgent
# Crawling for AI model training.

User-agent: Meta-ExternalFetcher
# Fetch for user requests. Can bypass robots.txt.
</code></pre>
<p><strong>Other relevant crawlers</strong>:</p>
<pre><code>User-agent: Amazonbot
User-agent: Bytespider      # ByteDance
User-agent: Applebot-Extended  # Apple AI (flag, not bot)
User-agent: CCBot           # Common Crawl
User-agent: cohere-ai
User-agent: cohere-training-data-crawler
</code></pre>
<h3>Configuration strategies</h3>
<p><strong>Strategy 1: Full access for maximum AI visibility</strong></p>
<pre><code># Allow all AI crawlers
User-agent: GPTBot
Allow: /

User-agent: ChatGPT-User
Allow: /

User-agent: OAI-SearchBot
Allow: /

User-agent: ClaudeBot
Allow: /

User-agent: anthropic-ai
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: Google-Extended
Allow: /

User-agent: Meta-ExternalAgent
Allow: /

User-agent: Amazonbot
Allow: /
</code></pre>
<p><strong>Strategy 2: AI search visibility, no training</strong></p>
<p>This configuration allows AI systems to cite your content in responses, but prevents use for training models:</p>
<pre><code># Allow search/user crawlers
User-agent: ChatGPT-User
Allow: /

User-agent: OAI-SearchBot
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: Perplexity-User
Allow: /

# Block training crawlers
User-agent: GPTBot
Disallow: /

User-agent: ClaudeBot
Disallow: /

User-agent: Google-Extended
Disallow: /

User-agent: Meta-ExternalAgent
Disallow: /

User-agent: CCBot
Disallow: /

User-agent: cohere-training-data-crawler
Disallow: /
</code></pre>
<p><strong>Strategy 3: Selective access by directory</strong></p>
<pre><code>User-agent: GPTBot
Allow: /blog/
Allow: /docs/
Disallow: /api/
Disallow: /internal/
Disallow: /user-data/
</code></pre>
<h3>Limitations of robots.txt</h3>
<p>A critical point: robots.txt is a voluntary protocol. Crawlers can ignore it.</p>
<p>In August 2025, <a href="https://www.medianama.com/2025/12/223-user-driven-ai-bots-crawling-grows-15x-in-2025-cloudflare-report/">Cloudflare blocked Perplexity bots</a> after documenting protocol violations. In October 2025, Reddit deliberately trapped Perplexity crawlers, demonstrating they bypassed restrictions through third-party tools. Legal action followed.</p>
<p>The operational consequence: robots.txt alone is not enough. For real enforcement, you need IP verification, WAF rules, or CDN-level blocks. Cloudflare reports that over 2.5 million sites use its managed robots.txt function to block AI training.</p>
<hr />
<h2>llms.txt: the new standard for guiding LLMs</h2>
<p>In September 2024, <a href="https://llmstxt.org/">Jeremy Howard of Answer AI proposed llms.txt</a>, a new standard file for communicating with Large Language Models. Unlike robots.txt, which controls access, llms.txt guides models toward the most relevant content.</p>
<h3>What llms.txt does</h3>
<p>The llms.txt file is a markdown document positioned at the domain root (<code>/llms.txt</code>). It works as a curated map that tells LLMs which pages contain the most important information and how to interpret them.</p>
<p>It’s not a blocking mechanism. It’s a recommendation system, like a librarian guiding a visitor to the right shelves instead of letting them wander.</p>
<h3>File structure</h3>
<pre><code class="language-markdown"># example.com

&gt; Technical site on AI implementations for enterprise.
&gt; Content verified, updated monthly.

## Core Documentation

- [Production RAG Guide](https://example.com/docs/rag-production):
  RAG architectures tested in production, chunking patterns,
  evaluation metrics. Updated Q4 2024.

- [API Reference](https://example.com/docs/api):
  Complete REST API documentation. Includes code examples
  in Python and cURL.

## Technical Articles

- [LLM Latency Optimization](https://example.com/blog/llm-latency):
  Strategies to reduce p95 latency below 200ms.
  Includes benchmarks on Claude, GPT-4, Mistral.

- [AI Cost Management](https://example.com/blog/ai-costs):
  Framework for estimating and optimizing inference costs.
  Real data from enterprise deployments.

## Resources

- [AI Glossary](https://example.com/glossary):
  Technical definitions of 150+ AI/ML terms.
</code></pre>
<h3>llms-full.txt: extended version</h3>
<p>Beyond llms.txt, the standard provides an optional <code>llms-full.txt</code> file containing the full site content in flattened format. It removes non-essential HTML, CSS, JavaScript and presents text only. Some sites generate files of 100K+ words.</p>
<p>The advantage: allows LLMs to process the entire site in a single context. The limitation: easily exceeds the context window of most models.</p>
<h3>Adoption status</h3>
<p>As of January 2025, <a href="https://www.ovrdrv.com/insights/llms-txt-the-new-standard-for-ai-crawling">OpenAI, Google, and Anthropic do not natively support llms.txt</a>. Their crawlers don’t automatically read the file.</p>
<p>Current adoption is concentrated in specific niches:</p>
<ul>
<li><strong>Technical documentation</strong>: Mintlify integrated llms.txt in November 2024. Documentation sites for Anthropic, Cursor, Cloudflare, Vercel use it.</li>
<li><strong>Dedicated directories</strong>: <a href="https://directory.llmstxt.cloud">directory.llmstxt.cloud</a> and <a href="https://llmstxt.site">llmstxt.site</a> catalog sites with implementations.</li>
<li><strong>Manual use</strong>: Developers who upload the file directly to ChatGPT or Claude to provide context.</li>
</ul>
<p>It’s an investment in future-proofing. When major providers adopt the standard, those who have already implemented will have an advantage.</p>
<h3>Implementation</h3>
<ol>
<li>Create <code>/llms.txt</code> at the domain root</li>
<li>UTF-8 format, clean markdown</li>
<li>Include only indexable pages (no noindex, no blocked in robots.txt)</li>
<li>Add concise but informative descriptions for each URL</li>
<li>Optional: reference in robots.txt with <code># LLM-policy: /llms.txt</code></li>
</ol>
<h3>Differences with other standard files</h3>
<table>
  <caption>Comparison of standard files for web and AI crawlers</caption>
  <thead>
    <tr>
      <th>File</th>
      <th>Purpose</th>
      <th>Target</th>
      <th>Format</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td>robots.txt</td>
      <td>Crawler access control</td>
      <td>Search engines, AI crawlers</td>
      <td>Plain text, directives</td>
    </tr>
    <tr>
      <td>sitemap.xml</td>
      <td>Complete page catalog</td>
      <td>Search engines</td>
      <td>XML</td>
    </tr>
    <tr>
      <td>llms.txt</td>
      <td>Curated priority content map</td>
      <td>LLM</td>
      <td>Markdown</td>
    </tr>
    <tr>
      <td>humans.txt</td>
      <td>Team credits</td>
      <td>Humans</td>
      <td>Plain text</td>
    </tr>
  </tbody>
</table>
<hr />
<h2>Structured Data and JSON-LD for AI</h2>
<p>Structured data is not new. It’s been standard SEO since 2011. But its role changes in the context of generative search engines.</p>
<h3>Why Structured Data matters for AI</h3>
<p>LLMs process everything as tokens. They don’t natively distinguish between a price, a name, a date. Structured data provides an explicit semantic layer that disambiguates content.</p>
<p>An article with JSON-LD markup communicates in a machine-readable way: this is the author, this is the publication date, this is the publishing organization, these are the sources cited. The model doesn’t have to infer this structure from the text.</p>
<h3>Basic JSON-LD implementation</h3>
<p>JSON-LD (JavaScript Object Notation for Linked Data) is the preferred format. It’s inserted in a <code>&lt;script&gt;</code> tag without mixing with content HTML:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "TechArticle",
  "@id": "https://example.com/rag-production-guide",
  "headline": "RAG in Production: Patterns and Anti-Patterns",
  "description": "Technical guide to enterprise RAG implementation with real metrics",
  "author": {
    "@type": "Person",
    "name": "Author Name",
    "url": "https://example.com/team/author-name",
    "jobTitle": "AI Team Leader",
    "knowsAbout": ["RAG", "LLM", "Vector Databases", "AI Engineering"]
  },
  "datePublished": "2025-01-04",
  "dateModified": "2025-01-04",
  "publisher": {
    "@type": "Organization",
    "name": "Example.com",
    "url": "https://example.com",
    "logo": {
      "@type": "ImageObject",
      "url": "https://example.com/logo.png"
    }
  },
  "mainEntityOfPage": {
    "@type": "WebPage",
    "@id": "https://example.com/rag-production-guide"
  },
  "articleSection": "Engineering",
  "keywords": ["RAG", "Production", "Enterprise AI", "Vector Search"],
  "wordCount": 3500,
  "inLanguage": "en"
}
&lt;/script&gt;
</code></pre>
<h3>Priority schema types for AI visibility</h3>
<p><strong>Article / TechArticle / NewsArticle</strong></p>
<p>For editorial content. TechArticle for technical documentation.</p>
<p><strong>FAQPage</strong></p>
<p>Q&amp;A structure that generative search engines can extract directly:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "What is the difference between SEO and GEO?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "SEO optimizes for lists of results from traditional search engines. GEO optimizes for being cited in synthesized responses from generative search engines like ChatGPT and Perplexity."
      }
    },
    {
      "@type": "Question",
      "name": "Does llms.txt replace robots.txt?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "No. robots.txt controls crawler access. llms.txt guides LLMs toward priority content. They have complementary functions."
      }
    }
  ]
}
&lt;/script&gt;
</code></pre>
<p><strong>HowTo</strong></p>
<p>For step-by-step guides:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "HowTo",
  "name": "How to configure robots.txt for AI crawlers",
  "step": [
    {
      "@type": "HowToStep",
      "position": 1,
      "name": "Identify target AI crawlers",
      "text": "Map the user-agents of AI crawlers you want to allow or block."
    },
    {
      "@type": "HowToStep",
      "position": 2,
      "name": "Define access strategy",
      "text": "Decide whether to allow training, search-only, or block completely."
    },
    {
      "@type": "HowToStep",
      "position": 3,
      "name": "Implement directives",
      "text": "Add User-agent and Allow/Disallow rules to your robots.txt file."
    }
  ]
}
&lt;/script&gt;
</code></pre>
<p><strong>Organization and Person: E-E-A-T for AI</strong></p>
<p>E-E-A-T (Experience, Expertise, Authoritativeness, Trustworthiness) is no longer just a Google framework. Data shows that LLMs verify author credentials before citing: 96% of content in AI Overview comes from sources with verified authors. Content with detailed author bios has 73% higher selection probability.</p>
<p>The Person schema must go beyond the name. It needs to communicate credentials, affiliations, specific expertise:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "Person",
  "name": "Mario Rossi",
  "url": "https://example.com/team/mario-rossi",
  "image": "https://example.com/images/mario-rossi.jpg",
  "jobTitle": "Senior AI Engineer",
  "description": "10+ years of experience in ML/AI, specialized in enterprise RAG systems",
  "worksFor": {
    "@type": "Organization",
    "name": "TechCorp Italia",
    "url": "https://techcorp.it"
  },
  "alumniOf": {
    "@type": "CollegeOrUniversity",
    "name": "Politecnico di Milano"
  },
  "hasCredential": [
    {
      "@type": "EducationalOccupationalCredential",
      "credentialCategory": "certification",
      "name": "AWS Machine Learning Specialty"
    },
    {
      "@type": "EducationalOccupationalCredential",
      "credentialCategory": "certification",
      "name": "Google Cloud Professional ML Engineer"
    }
  ],
  "knowsAbout": [
    "Retrieval-Augmented Generation",
    "Large Language Models",
    "Vector Databases",
    "MLOps",
    "AI Engineering"
  ],
  "sameAs": [
    "https://linkedin.com/in/mariorossi",
    "https://github.com/mariorossi",
    "https://scholar.google.com/citations?user=xxx"
  ]
}
&lt;/script&gt;
</code></pre>
<p><strong>E-E-A-T checklist for Person schema:</strong></p>
<ul>
<li><code>description</code> with years of experience and specialization</li>
<li><code>hasCredential</code> for verifiable certifications</li>
<li><code>knowsAbout</code> with specific topics (not generic)</li>
<li><code>sameAs</code> with links to verified profiles (LinkedIn, GitHub, Google Scholar)</li>
<li><code>alumniOf</code> for academic affiliations</li>
<li><code>worksFor</code> with organization URL</li>
</ul>
<h3>Citation schema</h3>
<p>For content that cites external sources, the Citation schema adds context:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "Article",
  "headline": "Analysis of GEO Paper Princeton",
  "citation": [
    {
      "@type": "ScholarlyArticle",
      "name": "GEO: Generative Engine Optimization",
      "author": ["Pranjal Aggarwal", "et al."],
      "datePublished": "2024",
      "publisher": {
        "@type": "Organization",
        "name": "Princeton University"
      },
      "url": "https://arxiv.org/abs/2311.09735"
    }
  ]
}
&lt;/script&gt;
</code></pre>
<h3>ImageObject and VideoObject for multimodal content</h3>
<p>Multimodal content has 156% higher probability of being selected in AI Overview compared to text-only content. Gemini and Perplexity invest heavily in multimodal search. Schema for media becomes relevant:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "ImageObject",
  "contentUrl": "https://example.com/images/architettura-rag.png",
  "name": "Enterprise RAG system architecture",
  "description": "Architectural diagram showing data flow between vector store, retriever and LLM in a production RAG system",
  "author": {
    "@type": "Person",
    "name": "Mario Rossi"
  },
  "datePublished": "2025-01-04",
  "encodingFormat": "image/png"
}
&lt;/script&gt;
</code></pre>
<p>For video with transcription:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "VideoObject",
  "name": "Deploy RAG in production: walkthrough",
  "description": "Video tutorial on deploying a RAG system on AWS with monitoring",
  "thumbnailUrl": "https://example.com/video/rag-deploy-thumb.jpg",
  "uploadDate": "2025-01-04",
  "duration": "PT12M30S",
  "transcript": "https://example.com/video/rag-deploy-transcript.txt",
  "author": {
    "@type": "Person",
    "name": "Mario Rossi"
  }
}
&lt;/script&gt;
</code></pre>
<p><strong>Best practices for AI-friendly media:</strong></p>
<ul>
<li>Descriptive and contextual alt text (not “image1.png”)</li>
<li>Captions that explain the content, not just describe it</li>
<li>Transcriptions for all videos</li>
<li>Captions that contextualize the figure in surrounding text</li>
</ul>
<h3>Real impact on AI search</h3>
<p>John Mueller of Google clarified in January 2025 that structured data is not a direct ranking factor. But the indirect impact is documented:</p>
<ul>
<li>Rich snippets from structured data increase CTR by 30% according to <a href="https://www.brightedge.com/">BrightEdge</a></li>
<li>72% of sites on Google’s first page use schema markup</li>
<li>Google’s AI Overviews process structured data to build responses</li>
</ul>
<p>Structured data doesn’t guarantee citations in generative search engines. But it provides the semantic context that facilitates correct interpretation of content.</p>
<hr />
<h2>Technical requirements for AI crawlers</h2>
<p>Beyond structured data, there are technical requirements that influence LLMs’ ability to process and cite content.</p>
<h3>Static HTML vs JavaScript rendering</h3>
<p>AI crawlers struggle with JavaScript-rendered content. Unlike Googlebot, which executes JS, many AI crawlers prefer or require static HTML.</p>
<p><strong>Operating rules:</strong></p>
<ul>
<li>Critical content must be present in static HTML, not generated dynamically</li>
<li>Avoid content hidden in tabs, accordions, or loaded on-scroll</li>
<li>If you use JS frameworks (React, Vue, Next.js), verify that SSR or SSG produces complete HTML</li>
<li>Test: view the page with JS disabled. What you see is what base AI crawlers see.</li>
</ul>
<h3>Content freshness signals</h3>
<p>23% of content selected in AI Overview is less than 30 days old. Perplexity indexes daily. Freshness signals are prioritized over historical authority.</p>
<p><strong>Implementation:</strong></p>
<p><code>dateModified</code> in schema must reflect actual updates:</p>
<pre><code class="language-html">&lt;script type="application/ld+json"&gt;
{
  "@context": "https://schema.org",
  "@type": "TechArticle",
  "headline": "Production RAG Guide",
  "datePublished": "2024-06-15",
  "dateModified": "2025-01-04"
}
&lt;/script&gt;
</code></pre>
<p><strong>Freshness checklist:</strong></p>
<ul>
<li>Update <code>dateModified</code> only for substantial changes (not typo fixes)</li>
<li>Prominently signal updates in content (“Updated: January 2025”)</li>
<li>Quarterly review of evergreen content</li>
<li>Update statistics and data at least annually</li>
<li>Remove or mark obsolete content as archived</li>
</ul>
<h3>Citation verification and fact-checking</h3>
<p>AI systems cross-reference with authoritative sources in real-time. Content with verifiable citations has 89% higher selection probability compared to content with unsupported claims.</p>
<p><strong>Rules:</strong></p>
<ul>
<li>Every statistic must have a linked source</li>
<li>“According to research” without a link = unverifiable claim = penalized</li>
<li>Prefer primary sources (papers, official documentation) over secondary sources</li>
<li>Citations from Wikipedia, Statista, Pew Research, arXiv papers carry more weight</li>
</ul>
<hr />
<h2>GEO strategies: what the research says</h2>
<p>The <a href="https://arxiv.org/abs/2311.09735">“GEO: Generative Engine Optimization” paper</a> from Princeton, Georgia Tech, Allen Institute, and IIT Delhi is the most rigorous available study on optimization for generative search engines. It tested 9 techniques on 10,000 queries.</p>
<h3>The three most effective strategies</h3>
<p><strong>1. Cite Sources: +40% visibility</strong></p>
<p>Adding citations from authoritative sources is the strategy with the highest overall impact. For sites with low ranking in traditional SERPs, the effect is even more pronounced: +115% for sites in fifth position.</p>
<p>Simply citing is not enough. The citation must be from a recognized, relevant source, verifiable.</p>
<p><strong>2. Quotation Addition</strong></p>
<p>Incorporating direct quotes from industry experts increases authenticity and perceived depth. Works particularly well for opinion-based content.</p>
<p><strong>3. Statistics Addition</strong></p>
<p>Quantitative data beats qualitative discussion. “42% of AI projects fail” has more impact than “many AI projects fail”. Works particularly well for Legal and Government domains.</p>
<h3>Structuring content for extraction: Answer Blocks</h3>
<p>LLMs don’t cite entire pages. They extract specific blocks. Optimizing for this pattern is critical.</p>
<p><em>Passage length</em> optimal: 134-167 words per citable block. For direct FAQ answers: 40-60 words. Content with summary boxes at the beginning has 28-40% higher citation probability.</p>
<p><strong>Practical implementation:</strong></p>
<ol>
<li>
<p><strong>TL;DR at the beginning:</strong> Every article opens with a self-contained summary block. It’s not just for human readers: it’s the block that LLMs preferentially extract.</p>
</li>
<li>
<p><strong>Self-contained sections:</strong> Each H2/H3 should be citable independently from the rest. An LLM should be able to extract that section and have a complete answer.</p>
</li>
<li>
<p><strong>Heading as questions:</strong> “What is RAG?” performs better than “RAG Overview”. Direct matching with conversational queries.</p>
</li>
<li>
<p><strong>Modular paragraphs:</strong> 75-300 words per section. No wall of text. Modular blocks are easier to extract and cite.</p>
</li>
<li>
<p><strong>Direct answers first, context after:</strong> The answer to the heading’s implicit question should appear in the first 2-3 sentences. Elaboration comes after.</p>
</li>
</ol>
<p><strong>Example of optimized structure:</strong></p>
<pre><code class="language-markdown">## What is the difference between SEO and GEO?

SEO optimizes for ranking in lists of results from traditional
search engines. GEO optimizes for being cited in synthesized
responses from generative search engines like ChatGPT, Perplexity
and Gemini. [40-60 words of direct answer]

The fundamental change concerns the objective: from ranking to
citation. In classical SEO, success is position 1 in the SERPs.
In GEO, success is being the source that the AI cites when responding.
[Elaboration and context]
</code></pre>
<h3>Domain-specific strategies</h3>
<p>The paper found that effectiveness varies by domain:</p>
<ul>
<li><strong>History</strong>: Authoritative and persuasive tone</li>
<li><strong>Facts</strong>: Citations from primary sources</li>
<li><strong>Law/Government</strong>: Statistics and quantitative data</li>
<li><strong>Science/Health</strong>: Technical terminology + authoritativeness</li>
</ul>
<h3>Platform-specific optimization</h3>
<p>Each LLM has different preferences. An effective GEO strategy considers these differences:</p>
<table>
  <caption>Optimization preferences for generative AI platforms</caption>
  <thead>
    <tr>
      <th>Platform</th>
      <th>Main preferences</th>
      <th>Optimization</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td><strong>ChatGPT</strong></td>
      <td>Wikipedia, popular brands, established content</td>
      <td>Authority building, Wikipedia presence if applicable</td>
    </tr>
    <tr>
      <td><strong>Perplexity</strong></td>
      <td>Reddit, recent content, real-time</td>
      <td>Freshness priority, community engagement</td>
    </tr>
    <tr>
      <td><strong>Gemini</strong></td>
      <td>Multimodal, Google ecosystem, schema markup</td>
      <td>Video, optimized images, complete structured data</td>
    </tr>
    <tr>
      <td><strong>Claude</strong></td>
      <td>Accuracy, balanced content, attribution</td>
      <td>Proper attribution, neutral and evidence-based framing</td>
    </tr>
    <tr>
      <td><strong>Google AI Overview</strong></td>
      <td>Top 10 organic, strong E-E-A-T</td>
      <td>Traditional SEO + extended structured data</td>
    </tr>
  </tbody>
</table>
<p><strong>Operational implications:</strong></p>
<ul>
<li>ChatGPT cites Wikipedia in 48% of responses. For topics with a Wikipedia entry, presence there matters.</li>
<li>Perplexity prefers Reddit (46.7% of citations). Content discussed in relevant subreddits has an advantage.</li>
<li>Gemini integrates images and video into responses. Multimodal content performs better.</li>
<li>Claude verifies accuracy more rigorously. Unsupported claims are discarded.</li>
</ul>
<h3>What doesn’t work</h3>
<p><strong>Keyword stuffing</strong>: Adding keywords from the query to content worsens visibility by 10% compared to baseline. Generative search engines penalize over-optimization.</p>
<p><strong>Generic persuasive language</strong>: Persuasive tone without substance doesn’t improve ranking.</p>
<h3>Democratization of results</h3>
<p>An interesting aspect: GEO levels the playing field. Sites with low ranking in traditional SERPs benefit more from GEO optimizations than dominant sites. Cite Sources brings +115% to sites in fifth position and -30% to sites in first position.</p>
<p>For small publishers and independent businesses, it’s an opportunity to compete with corporate giants without comparable SEO budgets.</p>
<hr />
<h2>Implementation checklist</h2>
<h3>robots.txt</h3>
<ul>
<li>[ ] Map all AI crawlers relevant to your industry</li>
<li>[ ] Define strategy: full access, search-only, selective</li>
<li>[ ] Implement directives for each user-agent</li>
<li>[ ] Verify syntax with <a href="https://www.google.com/webmasters/tools/robots-testing-tool">Google Robots Testing Tool</a></li>
<li>[ ] Monitor server logs for crawler activity</li>
<li>[ ] Verify actual compliance (IP check for suspicious crawlers)</li>
<li>[ ] Quarterly review: new crawlers emerge regularly</li>
</ul>
<h3>llms.txt</h3>
<ul>
<li>[ ] Create markdown file at domain root</li>
<li>[ ] Include site description and content type</li>
<li>[ ] Organize URLs by category/priority</li>
<li>[ ] Add concise descriptions for each link</li>
<li>[ ] Verify that all URLs are indexable</li>
<li>[ ] Consider llms-full.txt for sites with extended documentation</li>
<li>[ ] Update when new priority content is published</li>
</ul>
<h3>Structured Data / JSON-LD</h3>
<ul>
<li>[ ] Implement Organization schema for the site</li>
<li>[ ] Add Person schema for authors with complete E-E-A-T:
<ul>
<li>[ ] <code>description</code> with years of experience and specialization</li>
<li>[ ] <code>hasCredential</code> for verifiable certifications</li>
<li>[ ] <code>knowsAbout</code> with specific topics</li>
<li>[ ] <code>sameAs</code> with LinkedIn, GitHub, Google Scholar</li>
</ul>
</li>
<li>[ ] Use Article/TechArticle for editorial content</li>
<li>[ ] Implement FAQPage for Q&amp;A sections</li>
<li>[ ] Add Citation schema for research-based content</li>
<li>[ ] Implement ImageObject/VideoObject for media</li>
<li>[ ] Validate with <a href="https://search.google.com/test/rich-results">Google Rich Results Test</a></li>
<li>[ ] Verify parity between markup and visible content</li>
</ul>
<h3>GEO-optimized content</h3>
<ul>
<li>[ ] TL;DR of 40-60 words at the beginning of each article</li>
<li>[ ] Self-contained sections (citable independently)</li>
<li>[ ] Headings formulated as questions where appropriate</li>
<li>[ ] Modular paragraphs: 75-300 words per section</li>
<li>[ ] <em>Passage length</em>: 134-167 words for key blocks</li>
<li>[ ] Include citations from authoritative sources in every article</li>
<li>[ ] Add statistics and quantitative data with source</li>
<li>[ ] Use expert quotations where relevant</li>
<li>[ ] Avoid keyword stuffing</li>
<li>[ ] Calibrate tone for domain</li>
</ul>
<h3>Technical requirements</h3>
<ul>
<li>[ ] Critical content in static HTML (not JS-only rendering)</li>
<li>[ ] No content hidden in tabs/accordions/lazy-load</li>
<li>[ ] Test page with JavaScript disabled</li>
<li>[ ] <code>dateModified</code> updated for substantial changes</li>
<li>[ ] Signal updates in content (“Updated: Month Year”)</li>
<li>[ ] Quarterly review of evergreen content</li>
<li>[ ] Every statistic with linked source</li>
</ul>
<h3>Media and Multimodal</h3>
<ul>
<li>[ ] Descriptive and contextual alt text for images</li>
<li>[ ] Captions that explain the content</li>
<li>[ ] Transcriptions for all videos</li>
<li>[ ] ImageObject/VideoObject schema implemented</li>
<li>[ ] Captions that contextualize figures in surrounding text</li>
</ul>
<h3>Monitoring</h3>
<ul>
<li>[ ] Track AI crawler activity in server logs</li>
<li>[ ] Monitor brand mentions in ChatGPT/Perplexity/Gemini responses</li>
<li>[ ] Analyze competitor citation share</li>
<li>[ ] Measure referral traffic from AI platforms</li>
<li>[ ] Monthly metrics review</li>
</ul>
<hr />
<h2>The window of opportunity</h2>
<p>Cloudflare data shows that crawling for AI training still dominates traffic, with volumes 8 times higher than search crawling and 32 times higher than user-action crawling. But the trend is clear: user-action traffic is growing faster than any other category.</p>
<p>Those who implement GEO infrastructure now build advantages that accumulate over time. Citations generate other citations. Authority recognized by models strengthens. First-mover advantage in this space isn’t just about technical positioning: it’s about building an established presence before competition intensifies.</p>
<p>Traditional SEO doesn’t disappear. It continues to serve 70% of search traffic that still goes through classic SERPs. But the remaining 30%, and its growth trajectory, requires new tools.</p>
<hr />
<h2>Sources</h2>
<p>Aggarwal, P., et al. (2024). <a href="https://arxiv.org/abs/2311.09735"><em>GEO: Generative Engine Optimization</em></a>. arXiv:2311.09735. Princeton University, Georgia Tech, Allen Institute for AI, IIT Delhi.</p>
<p>AI Mode Boost. (2025). <a href="https://aimodeboost.com/resources/research/ai-overview-ranking-factors-2025/"><em>AI Overview Ranking Factors: 2025 Comprehensive Study</em></a>.</p>
<p>Cloudflare. (2025, December). <a href="https://blog.cloudflare.com/from-googlebot-to-gptbot-whos-crawling-your-site-in-2025/"><em>From Googlebot to GPTBot: Who’s Crawling Your Site in 2025</em></a>. Cloudflare Blog.</p>
<p>Dataslayer. (2025). <a href="https://www.dataslayer.ai/blog/google-ai-overviews-the-end-of-traditional-ctr-and-how-to-adapt-in-2025"><em>Google AI Overviews Impact 2025: CTR Down 61%</em></a>.</p>
<p>Howard, J. (2024, September). <a href="https://llmstxt.org/"><em>llms.txt Proposal</em></a>. Answer AI.</p>
<p>W3C Schema Community. (2024). <a href="https://schema.org/docs/documents.html"><em>Schema Vocabulary Documentation</em></a>.</p>
<p>SEO Sherpa. (2025, October). <a href="https://seosherpa.com/google-ai-search-guidelines/"><em>Google AI Search Guidelines 2025</em></a>.</p>
<p>Single Grain. (2025, October). <a href="https://www.singlegrain.com/search-everywhere-optimization/google-ai-overviews-the-ultimate-guide-to-ranking-in-2025/"><em>Google AI Overviews: The Ultimate Guide to Ranking in 2025</em></a>.</p>
<p>Yoast. (2025). <a href="https://yoast.com/structured-data-schema-ultimate-guide/"><em>Structured Data with Schema for Search and AI</em></a>.</p>
<p>Overdrive Interactive. (2025, July). <a href="https://www.ovrdrv.com/insights/llms-txt-the-new-standard-for-ai-crawling"><em>LLMs.txt: The New Standard for AI Crawling</em></a>.</p>
]]></content:encoded>
      <category>Engineering</category>
      <category>Business</category>
      <category>GEO</category>
      <category>SEO</category>
      <category>AI Search</category>
      <category>Schema.org</category>
      <category>robots.txt</category>
      <category>llms.txt</category>
      <atom:link rel="alternate" hreflang="it" href="https://ireneburresi.dev/blog/engineering/seo-vs-geo/"/>
    </item>
  </channel>
</rss>