• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar
  • Home
  • About Me
    • Teaching / Speaking / Events
  • Book: Governing AI
  • Book: Digital Factics X
  • AI – Artificial Intelligence
    • Ethics of AI Disclosure
  • AI Learning
    • AI Course Descriptions
  • AI Policy

@BasilPuglisi

Content & Strategy, Powered by Factics & AI, Since 2009

  • Headlines
  • My Story
    • Engagements & Moderating
  • AI Thought Leadership
  • Basil’s Brand Blog
  • Building Blocks by AI
  • Local Biz Tips

What 34 Reports Actually Told Us About AI: The Truth Behind the Hype, the Proof, and the Path Forward

March 4, 2026 by Basil Puglisi Leave a Comment

A synthesis of research from McKinsey, Google, OpenAI, Anthropic, BCG, IBM, Microsoft, WEF, Deloitte, OECD, the Future of Life Institute, and more, compiled and critiqued by a practitioner.

The Setup: Why This Matters More Than Another Hot Take

Alex Issakova curated and shared a collection of 34 leading AI research reports from the world’s most credible institutions, the majority published in 2025 and 2026 with several foundational reports from the preceding cycle included for continuity, and made them available to practitioners. This article builds on that collection and goes one step further: it doesn’t just surface what the reports said. It tells you what they agreed on, where they contradicted each other, where the vendor bias is visible, and, most critically, what any serious leader should actually do with these findings.

The starting premise is uncomfortable, and we should name it plainly before going further. We are living through one of the most consequential technology transitions in human history, and the boardroom conversations about it are largely built on second-hand opinions, vendor-produced statistics, and recycled enthusiasm. The reports in this collection represent the closest thing available to primary evidence. None of them are perfect, and several carry significant conflicts of interest, but read together they produce a macro picture that is both more promising and more alarming than the headlines suggest. What follows is our honest read of that picture.

Let us begin where the evidence begins: with failure.


The Failure Problem Nobody Wants to Name

The central statistical claim circulating across the practitioner literature is that up to 80% of AI projects fail, with some sources suggesting as many as 95% of generative AI pilots never reach production. Before building on those numbers, we owe them honest treatment.

Neither figure traces cleanly to a single, methodologically rigorous primary study; both represent directional consensus accumulated across multiple reports, practitioner surveys, and industry commentary rather than a unified measurement exercise. No research team followed a defined cohort of AI projects from inception to a clearly defined endpoint and arrived at 80% or 95% through controlled observation. The numbers circulate because they ring true to practitioners, not because any lab produced them, and that distinction matters. The underlying finding is not wrong; the specific percentages should simply be read as indicative signals rather than precise measurements. What the evidence base does support, clearly and repeatedly, is the directional claim: the majority of AI pilots that begin as experiments do not transition into sustained, value-generating production systems. That finding appears consistently across vendor reports, independent research, and practitioner assessments. The exact proportion is uncertain; the pattern is not.

There is also an important qualification that often gets lost in the failure conversation. Failure in this context rarely means the technology did not work. The models function, the APIs respond, and the demos impress; failure means that the strategy, the integration, the change management, the governance, and the business case were not sufficient to carry an experiment from pilot to sustained production deployment. We are not watching AI fail. We are watching organizations fail to deploy it.

Google Cloud’s compilation of 1,001 real-world generative AI use cases, updated to October 2025, provides the most direct counterevidence to blanket failure narratives. Google documented cases where Mercari projected a 500% ROI while reducing employee workloads by 20% — a projection, not yet a realized return at time of publication. AES, a global energy company, used AI agents built on Anthropic’s Claude models to automate energy safety audits, achieving a 99% reduction in audit costs, cutting audit time from 14 days to one hour, and improving accuracy by 10 to 20%. Fivecast, an OSINT security company, delivered a 400% ROI for intelligence analysts. These cases range from realized production outcomes to documented projections; together they represent the range of what disciplined deployment looks like.

The tension between the failure rate statistics and the documented success stories is not a contradiction but the central finding of the entire research library. AI delivers extraordinary returns for organizations that do it right, while the majority of organizations are not doing it right. The question every leader should be sitting with is why.

The research library treats failure frequency extensively but says less about failure cost, and that gap matters for boards more than it matters for practitioners. Two bounded cost frameworks give leaders the financial frame the directional statistics cannot. On the regulatory side, the EU AI Act establishes penalty tiers that function as board-level upper bounds: up to EUR 35 million or 7% of annual global turnover for prohibited practice violations, with additional tiers for other infringements. These are not hypothetical; they are enforcement-ready as of 2024. On the operational side, IBM’s Cost of a Data Breach 2025 provides defensible benchmark costs for security incidents, with shadow AI and weak governance controls documented as additive cost drivers that compound the base exposure. The practical method for any leadership team is to price expected loss for the three highest-risk AI deployments currently in operation or planned, using the regulatory maximum, the breach benchmark, and the organization’s own incident classification as inputs. Failure is not abstract once it carries a number that belongs on a balance sheet.

What went wrong in the failures: The evidence points consistently to four root causes. First, organizations treated AI as a technology project rather than a business transformation, assigning it to IT rather than to the business owners of the processes being changed. Second, they did not invest in data quality, integration architecture, or change management at a level commensurate with the ambition of the initiative. Third, they confused successful demos with viable production systems, under-estimating the gap between a convincing proof of concept and a robust, monitored, continuously maintained deployment. Fourth, and most importantly, they did not maintain meaningful human judgment in the loop, treating AI outputs as decisions rather than as inputs to decisions.

That fourth root cause is the one we want to spend a moment on, because it’s the one most likely to be rationalized away in the name of efficiency. The distinction between AI as input and AI as decision-maker is not a philosophical position; it’s an operational one with measurable consequences. One governance architecture developed to address this structurally is Checkpoint-Based Governance (CBG), developed through practitioner work rather than theoretical design, which holds this as its foundational principle: human judgment must structurally interrupt AI action before consequences occur. The principle didn’t originate in a whiteboard exercise; it emerged from watching what happens when that interruption is absent, and from a law enforcement operational background where checkpoint discipline around high-consequence decisions was standard practice, not optional policy. CBG is one practitioner’s response to this gap; the research supports the principle, not any single architecture.


What Is Actually Working and Where

Having sat with 1,001 documented production deployments, the most reliable indicator of successful AI deployment is specificity. Organizations that succeeded defined a narrow, high-frequency task, instrumented it carefully, and measured outcomes before expanding, while organizations that failed tried to transform broadly without anchoring to measurable value. That single pattern explains more of the variance than any other variable in the data.

Customer operations emerged as the most consistently successful deployment category. Commerzbank built a specialized chatbot handling over 2 million chats and successfully resolving 70% of all inquiries. DBS Bank reduced customer call handling times by 20%. NoBroker’s AI agents are projected to handle 25 to 40% of future customer calls. Wagestream handles more than 80% of internal customer inquiries with AI. The pattern is consistent across every one of these cases: high-volume, repetitive, information-retrieval tasks with clear resolution criteria are where AI delivers reliable, measurable value with manageable risk. If your organization hasn’t started here, this is where to start.

Knowledge work acceleration is the second major success category, and it’s the one we find most instructive for the governance conversation. BBVA’s 100,000 employees saved nearly three hours per week each by using AI to automate repetitive document tasks. Toyota reduced manual effort by over 10,000 person-hours per year through an AI-enabled manufacturing intelligence platform. Suzano reduced the time required for complex data queries by 95% among 50,000 employees. Deutsche Bank’s DB Lumina reduced the time for financial research from hours or days to minutes. In each case, the human remains the decision-maker while the AI accelerates the preparation and analysis. This is the human-AI collaboration model working as designed. Within the HAIA-RECCLIN framework, one practitioner’s structured approach to this model, it maps directly to the Researcher role: AI accelerates evidence gathering and synthesis while the human arbiter exercises judgment on what the evidence means and what to do with it. The Deutsche Bank model isn’t a novel discovery; it is an independently validated confirmation of the underlying principle, arriving from a completely different operational context, and other governance architectures map to the same pattern.

Code generation and developer productivity is the third high-performing category, though it carries important caveats we should not gloss over. Wayfair developers with access to AI code assistance set up environments 55% faster, saw a 48% increase in code performance during unit testing, and 60% reported greater focus on more satisfying work. CME Group developers using AI code assistance reported productivity gains of at least 10.5 hours per month. Capgemini observed workload gains and more stable code quality across software engineering processes. The caveat is that code generation is also the category most prone to unchecked quality problems when human review is reduced or eliminated. The productivity gains are real; the governance requirement doesn’t disappear because the output is code rather than a decision.

Supply chain and demand forecasting delivered some of the most technically impressive results in the entire library. OTTO improved demand forecasting accuracy by up to 30%, reducing inventory costs. Coop achieved a 43% improvement in forecasting accuracy, reducing food waste. Simbe’s retail intelligence platform delivered a 4x return on investment within 90 days through AI-powered inventory monitoring. These outcomes reflect the fundamental strength of machine learning on large, structured, historical datasets where patterns are consistent and outcomes are measurable. This is AI operating in its native environment.

A note on the open-source pathway: The success cases documented across this library are drawn predominantly from deployments of proprietary frontier models, and that weighting reflects the source library, not the full deployment picture. According to Menlo Ventures’ 2025 enterprise generative AI survey, open-weight models including Meta’s Llama series remain the most widely adopted open-weight option in enterprise contexts even as their overall share of new deployments has shifted, and the Stanford HAI 2025 AI Index documents a meaningful rise in the share of open-weight foundation model releases alongside a narrowing capability gap on several benchmarks. For practitioners, open-weight deployment represents a different risk surface rather than a safer or more dangerous one: it reduces vendor lock-in and data residency risk while introducing model provenance, versioning, and internal governance requirements that proprietary API deployments handle externally. Organizations for whom data sovereignty, cost control, or vendor concentration risk are strategic concerns should treat open-weight deployment as a parallel pathway with its own governance requirements, not as a workaround to the governance principles this synthesis documents. The failure modes and success patterns are structurally similar; the accountability architecture has to be built internally rather than inherited from a provider.


The Macro Numbers Every Leader Needs

We want to be careful here not to let the success cases set the wrong baseline, because the macro picture is more sobering than the headline deployments suggest.

The OECD’s progress report on the European Union’s Coordinated Plan on Artificial Intelligence, published in 2025, provides the most credible macro baseline against excessive optimism. In 2024, approximately 13.5% of EU enterprises reported using AI technologies, up from around 8% the previous year, which represents meaningful growth; it also means that 86.5% of European enterprises had not yet operationalized AI technology by 2024. We are not in an era of broad AI adoption. We are in an era of early movers pulling ahead while the majority of organizations are still watching.

The OECD also documents a structural problem that no individual organization can solve alone: the AI skills gap. Germany and France are the strongest in Europe for AI engineering skills penetration, but most EU member states remain below the global average, and several countries, including Greece, Hungary, and Italy, are net exporters of AI talent, losing skilled professionals faster than they’re trained. This is a fundamental constraint on adoption at scale, not a minor friction, and it doesn’t resolve itself without deliberate investment.

Infrastructure is the second structural constraint, and the numbers here are genuinely striking. The OECD projects that demand for data centre capacity could triple by 2030, growing between 19 and 27% annually, and by that time generative AI alone is expected to account for approximately 40% of total AI-related infrastructure demand. Meeting this demand will require building at least twice the data centre capacity constructed globally since the year 2000. In 2024, more than 90% of global venture capital investment in AI, amounting to approximately 137 billion euros, was directed toward AI start-ups in the United States, while EU start-ups attracted only 9.15 billion euros. The competitive gap is structural and growing, and no amount of strategy announcements closes an infrastructure gap of that magnitude.

The productivity potential remains enormous, and this is the finding that justifies the urgency. McKinsey’s research on the economic potential of generative AI estimates that it could add the equivalent of 2.6 to 4.4 trillion dollars annually across 63 use cases analyzed. The majority of that value sits in four domains: customer operations, marketing and sales, software engineering, and research and development, which are exactly the domains where the documented success cases in the Google Cloud report are concentrated. That cross-source alignment is meaningful. When independent research organizations converge on the same value zones from different methodological starting points, the directional signal is worth trusting.


The Agentic Shift and Why 2026 Is Different

Something changed in the 2025 and 2026 research cycle that deserves more attention than it’s getting in most strategy conversations. The acceleration of agentic AI is not a forecast anymore; it’s a present-tense operational reality, and the governance implications haven’t caught up.

Until recently, enterprise AI was primarily reactive: a human posed a question or submitted a document, the AI responded, and a human reviewed the result. Agentic AI changes this model fundamentally. Agents take sequences of actions, use tools, access external systems, make intermediate decisions, and produce outcomes rather than just outputs. The human is no longer in the loop at every step by default; the human has to design the loop with that oversight built in.

The Anthropic 2026 State of AI Agents Report and the Deloitte enterprise AI reports both signal that agentic deployment is no longer a future capability; it is present reality in leading organizations. The documented evidence from the Google Cloud use cases supports this: spring.new customers achieve 95 to 99% time savings on R&D projects, with users creating applications in one to two hours that previously required three months. Guane’s AI platform reduced land restitution sentence processing from 12 hours to 6 minutes. Replit Agent, powered by Claude 3.5 Sonnet, has enabled over 100,000 application deployments, including projects that have grown into companies worth tens of millions of dollars.

The governance implications of this shift are profound, and we don’t think most organizations have fully reckoned with them. When AI functions as a response generator, human review of every output is feasible. When AI is an agent taking chains of actions, the oversight model must change fundamentally, shifting the governing question from “did we review this output?” to “did we design the system with appropriate checkpoints, logging, human escalation triggers, and rollback capabilities?” Most organizations deploying agentic AI today are not yet asking these questions systematically, and that gap is where the next generation of expensive failures will originate.

One governance infrastructure specification designed to answer these questions at scale is the GOPEL specification (Governance Orchestrator Policy Enforcement Layer), one practitioner’s architecture for this problem. GOPEL performs zero cognitive work by design: it dispatches, collects, routes, logs, pauses, hashes, and reports, seven deterministic operations and no others. The non-cognitive design is a deliberate security decision, because a governance layer that can evaluate content can also be manipulated, while a deterministic pipe reduces that attack surface significantly. Whether organizations adopt GOPEL or another architecture entirely, the principle the research library supports is clear: agentic systems require governance infrastructure built in before deployment, not retrofitted after incidents accumulate.

IBM’s Enterprise in 2030 report and their work on agentic AI’s strategic ascent both emphasize that the organizations defining competitive advantage in the next five years are those building agentic infrastructure now, not as an experiment, but as a production capability with governance architecture to match. We think that framing is right, and we think most organizations are running about two years behind where they need to be on it.


The Safety Crisis the Boardroom Refuses to Take Seriously

No section of this synthesis is more important, or more likely to be skipped, than this one. We’re going to ask you to stay with it.

The Future of Life Institute’s AI Safety Index, Winter 2025 edition, evaluated eight leading AI companies across 35 indicators spanning six critical domains. The findings are alarming in their directness, and they should be read as such. Anthropic received a C+, OpenAI received a C+, Google DeepMind received a C, and every other company evaluated received a D or below. No company scored above a D in the existential safety domain, and this was the second consecutive edition with that outcome. Read that again: the second consecutive edition.

The FLI’s independent review panel, comprising AI researchers from MIT, UC Berkeley, the University of Wisconsin-Madison, and other institutions, found that frontier AI companies’ safety commitment continues to lag far behind capability ambition, and that even the strongest performers lack the concrete safeguards, independent oversight, and credible long-term risk-management strategies that such powerful systems demand. This is the independent assessment of credentialed researchers reviewing publicly available evidence and company-provided surveys, not fringe concern from AI safety advocates. The gap between capability advancement and safety infrastructure is structural and widening, and no one inside the industry has a fully credible plan for closing it.

We want to spend a moment on Anthropic specifically, because the start of 2026 produced something practitioners should study carefully. In January 2026, Anthropic published Claude’s Constitution, an approximately 80-page document articulating values, character formation, and behavioral guidelines for its AI system. Days later, CEO Dario Amodei published a 20,000-word essay on AI risk that explicitly called for external legislation and societal response. Anthropic holds approximately 32% of enterprise LLM market share. These were not casual publications; they were a coordinated statement from the most safety-conscious major AI lab in operation.

The documents are genuinely impressive. Taken together, they represent comprehensive Ethical AI coverage, addressing character formation, honesty properties, corrigibility, and internal safeguards, and serious Responsible AI coverage, addressing training methodology, staged autonomy, and internal principal hierarchy. Anthropic deserves credit for the transparency and rigor of both. We’ve read them closely and mean that assessment sincerely.

And yet the FLI grade stands: C+. Understanding why requires a distinction that the field collapses at its own cost. There are three different categories operating under the broad label of AI safety. Ethical AI asks whether something should be done, and Anthropic’s Constitution addresses this comprehensively. Responsible AI asks who answers when it fails, with a ceiling at the absence of individual human oversight, and Anthropic addresses this seriously. AI governance asks who decides, by what authority, at what checkpoint, with external mechanisms that survive commercial pressure, and neither document provides that architecture. Amodei himself calls for external legislation in the essay, which is an acknowledgment that the governance layer he’s describing isn’t something Anthropic can build for itself. Internal governance, however sophisticated, remains voluntary. External governance begins when someone outside the builder can stop the system and that stop survives commercial pressure.

The practical consequence for enterprise leaders is what we’d call confidence without control. The sophistication of Anthropic’s published positions inspires confidence, and it should; the philosophical depth and transparency are real. But confidence is not control, and enterprise compliance requirements in healthcare, financial services, and government demand documented human oversight with audit trails traceable to named individuals with stop authority. Claude’s Constitution shapes Claude’s character. It doesn’t specify the checkpoint architecture that enterprise governance requires. Every organization deploying these systems needs to build that layer themselves, and most of them haven’t, because reading the Constitution felt like enough.

The best-governed AI company in the field earned a C+ from independent evaluators. Corporate AI governance is almost universally less developed than frontier AI company governance. The organizations treating a character document as a governance substitute are carrying a gap they haven’t yet priced.

Watermarking and content provenance remain a parallel concern worth naming. Chinese-regulated AI companies comply with binding national standards requiring both explicit and implicit watermarking, while major Western providers do not yet enforce equivalent standards consistently. As AI-generated content proliferates, the absence of provenance infrastructure creates escalating risks that most enterprise risk registers have not yet accounted for.


What the Board Needs to Own

The research on board readiness is the section of this synthesis that we find most personally frustrating, because the gap between what boards need to do and what boards are actually doing is wide and closing slowly.

The KPMG Boardroom View on Gen AI Adoption and McKinsey’s analysis of how boards can evolve on AI governance are united in a central finding: most boards are not equipped to govern AI effectively, and most are not moving fast enough to become equipped. The Australian Institute of Company Directors’ research on AI use by directors and boards, and the Institute of Directors’ work on AI governance in the boardroom, identify the same structural gap. Directors understand that AI is strategically important, but they don’t feel confident evaluating AI risk and are deferring to management on AI strategy while simultaneously being expected to provide oversight of it. That is not a tenable governance position, and the organizations that treat it as one are accumulating risk they don’t know they’re carrying.

The research evidence points to three things boards must own directly, rather than delegate, and we want to be precise about what ownership means in each case.

The first is AI risk classification. Not all AI risk is equal. Deploying AI to generate marketing copy carries fundamentally different risk than deploying AI agents to make credit decisions, process legal documents, or operate in regulated healthcare environments, and boards need a risk taxonomy that distinguishes these categories and assigns appropriate governance requirements to each. This maps to what CBG calls risk-proportional implementation: low-stakes contexts may consolidate checkpoints while maintaining audit trails, while high-consequence decisions require multiple checkpoints with independent review and distributed authority. Without this taxonomy, boards cannot tell whether the AI governance they’re being briefed on is appropriate to the risk they’re actually carrying.

The second is accountability architecture. When an AI system produces a harmful outcome, it must be possible to identify which human was responsible for the decision to deploy that system in that context with those constraints, which requires explicit accountability mapping before deployment rather than after an incident. If that mapping doesn’t exist in your organization today, it needs to exist before the next significant deployment.

The third is ongoing monitoring. AI systems are not static after deployment; model behavior can drift, data distributions change, and new use cases emerge that were not anticipated at the outset. Boards need assurance that management has continuous monitoring in place, not just initial deployment reviews. A deployment sign-off without a monitoring commitment is not governance.

McKinsey’s research on the agentic organization argues that the next paradigm for the AI era requires boards to reconceptualize what oversight means when autonomous systems are making sequences of consequential decisions at machine speed, because the governance frameworks built for human decision-making are insufficient for the task. We agree, and we’d add that the reconceptualization needs to happen before the agentic deployments scale, not after.


The Geography of Competitive Advantage

Geography matters more in the AI race than most enterprise strategy conversations acknowledge, and the numbers here are worth sitting with before we move to prescriptions.

The OECD report provides the most reliable picture of AI competitive advantage by country. The United States dominates AI venture capital investment, holding more than 90% of global flows in 2024. China has surpassed both the EU and the United States in total AI research publications since 2019 and is rapidly closing the gap in high-impact publications. The EU ranks second in high-impact AI research, slightly ahead of the United States, but significantly trails in commercialization, start-up formation, and infrastructure investment. Research leadership and deployment leadership are not the same thing, and the EU is learning that distinction at significant competitive cost.

For corporate leaders outside the United States, this geography is a strategic constraint, not an abstraction. European organizations face higher regulatory compliance costs under the EU AI Act, more limited access to frontier AI infrastructure, a smaller domestic talent pool, and a start-up ecosystem that is underfunded relative to the opportunity. The European Union’s 9.15 billion euros in AI venture capital against the United States’ 137 billion euros in 2024 represents a structural gap that no national AI strategy can close in the near term, and the organizations operating in that environment need to build their AI strategies with that constraint explicitly accounted for.

The OECD productivity research on the AI divide further complicates the picture in ways that should concern any leader who assumed AI would be a great equalizer. AI is not a neutral productivity enhancer that lifts all boats; it amplifies the advantages of organizations and countries that already have strong data infrastructure, technical talent, and capital access. The risk is not just that some organizations fall behind but that the gap compounds, which means the window for building a credible position is narrowing, not expanding.

OECD analysis shows that AI is reshaping job roles for most professions rather than replacing them entirely, and that framing is more accurate than the binary replacement narrative. Customer service workers now use AI-powered tools to manage inquiries while manufacturing employees engage with automated systems to improve efficiency. But accuracy on the framing doesn’t eliminate the workforce disruption challenge. Upskilling and reskilling initiatives are critical, and more than half of EU member states have launched broader national digital skills strategies, but AI-specific training remains less common and less systematic than the challenge demands.


The Things You Should Not Do

The research evidence, taken as a whole, produces a clear list of things organizations should actively avoid. These are not theoretical cautions assembled from first principles; they are patterns that appear repeatedly in the documented failure modes, and they carry the weight of that evidence.

Do not conflate pilot success with production readiness. The gap between a successful pilot and a sustainable production deployment is wider for AI than for almost any other technology category. Production AI requires monitoring infrastructure, feedback loops, model versioning, incident response capabilities, and change management at a scale that most organizations don’t build into their initial business cases. The pilot felt like success. The production deployment is where the real test begins.

Do not deploy AI without explicit accountability mapping. Every AI deployment should have a named human accountable for the outcomes it produces. When no human is accountable for an AI’s decisions, no human is positioned to catch its errors, which makes this the minimum requirement for responsible deployment rather than a bureaucratic compliance exercise. If you can’t name that human before go-live, you aren’t ready to go live.

Do not treat AI outputs as decisions. This applies at every level of the organization. An AI recommendation is an input to a human decision, not a substitute for it. Organizations that remove human judgment from consequential decisions in the name of efficiency are trading short-term cost reduction for long-term liability and error risk, and the research documents that trade-off consistently and unfavorably.

Do not underinvest in data quality. The performance of every AI system is bounded by the quality of the data on which it was trained and on which it operates. Organizations that deploy AI on poor-quality, incomplete, or biased data will produce poor-quality, incomplete, or biased outcomes at scale. This is a data governance problem, not a technology one, and no amount of model sophistication compensates for it.

Do not mistake vendor metrics for independent evidence. A significant portion of the AI research library is produced by organizations with direct financial interests in the conclusions. Google’s 1,001 use cases document is a marketing document as much as it is a research document; it contains genuine value and documented outcomes, but it presents those outcomes without reference to the initiatives that failed on the same platforms. We read it carefully and found it genuinely useful, and we also read it knowing what it was designed to do. Multi-AI review processes, where independent AI platforms evaluate the same evidence set and their convergence and divergence is documented, help surface what single-source analysis obscures.

Do not assume your AI governance framework is adequate. The Future of Life Institute’s finding that even the best-governed AI companies received grades no higher than C+ should be sobering for enterprise leaders. Corporate AI governance is almost universally less developed than frontier AI company governance, so enterprise organizations deploying those same systems at scale need to compensate with their own oversight architecture. If the organizations building the models are still working to govern them at a C+ level, the organizations deploying them need to take that reality seriously.


The Things You Should Do, With Evidence

Anchor every AI initiative to a measurable business outcome before you begin. The most successful documented deployments in this library share a common structure: a specific process, a defined metric, a baseline measurement, a deployment, and a post-deployment measurement. AES didn’t deploy AI broadly across energy operations; it deployed AI specifically to energy safety audits and measured audit cost, audit time, and audit accuracy, and that specificity enabled both the success and the measurement. This structure, a verified fact paired with a concrete tactic anchored to a measurable KPI, is a widely documented best practice that one practitioner’s methodology, Factics, has formalized since 2012. The research library validates the underlying principle independently across thirty-four separate studies and from completely different organizational contexts. Factics didn’t invent the idea of measurement; the research confirms it matters, consistently, at scale, and that confirmation holds regardless of which methodology an organization uses to operationalize it.

Build for human-AI collaboration, not for human replacement. The preponderance of evidence across the research library supports a model in which AI handles high-frequency, data-intensive tasks while humans retain judgment on consequential decisions. This isn’t a philosophical position; it is an empirically supported performance model that the research validates across industries, geographies, and organizational sizes. Deutsche Bank’s financial analysts using AI produce research faster, and they produce better-reviewed, more defensible research because a human analyst is still exercising professional judgment on the AI-accelerated output. The human didn’t get replaced; the human got better.

Invest in AI literacy at every level, including the board. The OECD data on skills gaps and the governance research on board readiness converge on the same prescription. Organizations that build competitive AI advantages are those where AI literacy extends from the technical team to the executive suite to the board, and the gap between where most boards currently sit and where they need to be is wide enough to constitute a material governance risk. This doesn’t require every leader to understand transformer architectures; it requires every leader to understand how to evaluate AI risk, AI value claims, and AI governance requirements. The AI Skills Initiative investments documented across the EU member states, from Austria’s 6 million euro Digital Skills Initiative to France’s 560 million euro AI Cluster initiative, reflect a recognition that skills investment isn’t optional, even when it’s expensive.

Design governance before you need it. The research evidence is consistent that organizations which try to retrofit governance onto deployed AI systems face significantly higher remediation costs and reputational risks than organizations that design governance architecture into the deployment from the beginning. One practitioner response to this finding is CBG’s Phase 0 recommendation: adopt manual governance before building automated systems, because the discipline of structured human oversight must precede the infrastructure. The principle applies regardless of which governance framework an organization selects. The governance architecture should address who approves deployment, what monitoring is required, what triggers human review or system rollback, and who is accountable for outcomes. Answering those questions before deployment isn’t bureaucracy; it’s the minimum viable checkpoint architecture, and the organizations that skip it are the ones generating the failure statistics in Part One.

Treat AI safety as a business continuity issue, not a philosophical debate. The Future of Life Institute’s safety index findings, the Anthropic CEO’s public warnings, and the emerging regulatory environment represented by the EU AI Act and California’s SB 53 are not abstract threats. They are signals that the operating environment for AI deployment is tightening, and organizations that build safety and compliance into their AI programs now will face lower regulatory risk and lower remediation costs than those that treat safety as someone else’s problem. It isn’t someone else’s problem; it’s the deployment environment every enterprise is operating inside.

Start with use cases where failure is recoverable. A generative AI tool that produces a poorly written marketing email is a recoverable failure. An AI agent that makes an incorrect credit decision affecting a customer’s financial stability is a more serious one. An AI system operating in a medical context that misclassifies a condition is potentially catastrophic. Start with high-volume, recoverable-failure use cases, build your monitoring and governance infrastructure, and earn the right to deploy in higher-consequence contexts by showing operational discipline in lower-consequence ones first. The organizations that built the documented success cases in this library almost all followed this sequence, even when they didn’t call it that.


The Honest Macro Picture

Having worked through 34 reports across ten domains of evidence, here is the macro picture as clearly as we can state it.

AI delivers real, measurable, documented business value now, at production scale, in hundreds of organizations across every major industry. The value is concentrated in high-frequency, data-intensive tasks, customer operations, knowledge work acceleration, and software development, and the productivity gains are real and significant. The success cases in this library aren’t cherry-picked anomalies; they represent a replicable pattern for organizations willing to do the structural work.

AI is also failing at scale, and we should not let the success cases obscure that. The majority of AI pilots do not reach production, and the majority of organizations that attempt deployment do not do it effectively; the failures are not primarily technology failures but strategy, governance, and change management failures. That diagnosis matters because it points to solvable problems, not fundamental limitations.

The safety infrastructure for frontier AI systems is inadequate relative to the capabilities being deployed. The best-governed AI companies in the world received grades no higher than C+ from independent evaluators, and the gap between capability advancement and safety preparedness is structural and widening. No organization deploying AI, including the organizations building the foundational models, has fully solved this problem, and the organizations pretending otherwise are carrying risk they haven’t priced.

The competitive geography is hardening faster than most enterprise strategy conversations acknowledge. The United States and China are building structural advantages in AI investment, talent, and infrastructure that will compound over time, and the organizations that will sustain competitive advantage are those that build genuine operational capability now, not those that make the most impressive announcements later.

The workforce transition is real but manageable with investment. For most professions, AI is reshaping roles rather than eliminating them, and the organizations and governments investing systematically in skills development now are building human capital that will determine competitive position through the decade. The ones waiting for the transition to stabilize before investing will find the gap has already compounded.


The Decision Point

The evidence from 34 of the world’s most rigorous AI research institutions, stripped of vendor enthusiasm and alarmist overcorrection, resolves into a single strategic imperative: organizations that build disciplined, governance-anchored, human-centered AI capabilities now will create advantages that compound, while organizations that continue to treat AI as a series of exciting pilots without the infrastructure to scale and the accountability structures to govern will continue to contribute to the majority-failure pattern the research documents consistently.

The failure isn’t inevitable, and the evidence for that claim sits in the same reports; the organizations beating the odds are not exceptional organizations but disciplined ones, and that distinction is what this synthesis is ultimately about.

The question is not whether AI is real. The question is whether your organization is.


FAQ – Frequently Asked Questions

Q: Why do most AI projects fail? Most AI projects fail because of governance, strategy, and change management gaps — not because the technology stopped working. Organizations treat AI as an IT project rather than a business transformation, skip data quality investment, and remove human judgment from consequential decisions. The models function. The deployment infrastructure around them does not.

Q: What AI deployments are actually delivering results in 2025 and 2026? Customer operations, knowledge work acceleration, code generation, and supply chain forecasting are producing the most consistent, measurable returns. Organizations that defined a narrow task, instrumented it carefully, and measured outcomes before expanding are generating documented ROI. Broad transformation attempts without anchored measurement continue to fail at the majority rate the research documents.

Q: What did the Future of Life Institute’s AI Safety Index find? The FLI’s Winter 2025 AI Safety Index found that the best-governed AI companies in the world, including Anthropic and OpenAI, scored no higher than C+. No company scored above a D in the existential safety domain, marking the second consecutive edition with that outcome. The gap between capability advancement and safety infrastructure is structural and widening.

Q: What is the difference between ethical AI, responsible AI, and AI governance? Ethical AI asks whether something should be done. Responsible AI asks who answers when it fails, with a ceiling at the absence of individual human oversight. AI governance asks who decides, by what authority, at what checkpoint, with external mechanisms that survive commercial pressure. Most organizations have the first and are developing the second; the third remains the critical gap.

Q: What should boards own directly on AI? Boards must directly own three things: AI risk classification that distinguishes marketing copy generation from agentic credit decisions, accountability architecture that maps a named human to every deployment before an incident occurs, and ongoing monitoring assurance that covers model drift, data distribution changes, and emerging use cases. Deferring all three to management is not a tenable governance position.

Q: Is open-source AI a viable enterprise option? Open-weight models including Meta’s Llama series represent a strategically viable pathway for organizations with data sovereignty, cost control, or vendor concentration concerns. Enterprise adoption remains significant. The governance requirements differ from proprietary deployment rather than disappear: model provenance, versioning, and internal oversight architecture must be built internally rather than inherited from a provider.

Q: What does AI failure actually cost an organization? The EU AI Act establishes regulatory exposure up to EUR 35 million or 7% of annual global turnover for prohibited practice violations. IBM’s Cost of a Data Breach 2025 benchmarks operational exposure, with shadow AI and weak governance controls documented as additive cost drivers. Organizations should price expected loss for their three highest-risk deployments before the next governance conversation.

Q: What makes agentic AI governance different from standard AI deployment governance? Agentic AI takes sequences of actions, accesses external systems, and produces outcomes rather than just outputs. Human review of every output is no longer the operational model. The governing question shifts from reviewing outputs to designing systems with appropriate checkpoints, logging, human escalation triggers, and rollback capabilities before deployment, not after incidents accumulate.


Sources

Primary Library (2025–2026)

Google Cloud. 1,001 Real-World Gen AI Use Cases from the World’s Leading Organizations. October 2025. https://cloud.google.com/transform/101-real-world-generative-ai-use-cases-from-industry-leaders

Future of Life Institute. AI Safety Index Report: Winter 2025. December 2025. Landing page: https://futureoflife.org/ai-safety-index-winter-2025/ Full PDF: https://futureoflife.org/wp-content/uploads/2025/12/AI-Safety-Index-Report_131225_Full_Report_Digital.pdf

OECD. Progress in Implementing the European Union Coordinated Plan on Artificial Intelligence, Volume 1. 2025. https://www.oecd.org/en/publications/oecd-reviews-of-digital-transformation-progress-in-implementing-the-european-union-coordinated-plan-on-artificial-intelligence-2025_5d8b78a9-en.html (Note: If this URL returns a 404, search oecd.org for “EU Coordinated Plan AI 2025” — OECD periodically restructures document URLs.)

Anthropic. 2026 State of AI Agents Report. 2026. https://www.anthropic.com/research/state-of-ai-agents (Verify current URL — Anthropic restructures research pages. If unavailable, search anthropic.com for “State of AI Agents 2026”.)

McKinsey Global Institute. The State of AI in 2025: Agents, Innovation, and Transformation. November 2025. https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai PDF: https://www.mckinsey.com/~/media/mckinsey/business%20functions/quantumblack/our%20insights/the%20state%20of%20ai/november%202025/the-state-of-ai-2025-agents-innovation_cmyk-v1.pdf

McKinsey Global Institute. The Economic Potential of Generative AI: The Next Productivity Frontier. June 2023. https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/the-economic-potential-of-generative-ai-the-next-productivity-frontier [Foundational valuation framework; included for cross-library continuity. Published outside the 2025–2026 primary window.]

McKinsey. The AI Reckoning: How Boards Can Evolve. December 2025. https://www.mckinsey.com/capabilities/mckinsey-technology/our-insights/the-ai-reckoning-how-boards-can-evolve

McKinsey. Superagency in the Workplace: Empowering People to Unlock AI’s Full Potential. 2025. https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/superagency-in-the-workplace-empowering-people-to-unlock-ais-full-potential

McKinsey. The Agentic Organization: Contours of the Next Paradigm for the AI Era. 2025. https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/the-agentic-organization

McKinsey. McKinsey Global Tech Agenda 2026. 2026. https://www.mckinsey.com/capabilities/mckinsey-technology/our-insights/the-mckinsey-technology-trends-outlook (Search mckinsey.com for “Global Tech Agenda 2026” if URL resolves to the general trends page.)

BCG. AI Radar 2026: As AI Investments Surge, CEOs Take the Lead. January 2026. PDF: https://web-assets.bcg.com/73/8e/cc44cbc14a3b81695f8a3de28ff1/ai-radar-2026-web-jan-2026-edit.pdf Landing page: https://www.bcg.com/publications/2026/ai-radar-2026

BCG. Targets Over Tools: The Mandate for AI Transformation. 2025. https://www.bcg.com/publications/2025/targets-over-tools-the-mandate-for-ai-transformation

BCG. The State of Gen AI in Global Financial Institutions. 2025. https://www.bcg.com/publications/2025/state-of-gen-ai-in-global-financial-institutions (Search bcg.com for “state of gen AI financial institutions 2025” if URL redirects.)

IBM Institute for Business Value. The Enterprise in 2030. 2025. https://www.ibm.com/thought-leadership/institute-business-value/en-us/report/enterprise-2030

IBM Institute for Business Value. Agentic AI’s Strategic Ascent. October 2025. https://www.ibm.com/thought-leadership/institute-business-value/en-us/report/agentic-ai-operating-model

Deloitte AI Institute. Agentic Enterprise 2028: A Blueprint for Growth. 2025. https://www.deloitte.com/us/en/what-we-do/capabilities/applied-artificial-intelligence/articles/agentic-ai-enterprise-2028.html

Deloitte AI Institute. State of AI in the Enterprise 2026 (Now Decides Next). January 2026. https://www.deloitte.com/us/en/what-we-do/capabilities/applied-artificial-intelligence/content/state-of-ai-in-the-enterprise.html

Deloitte. NL Gen AI ROI Survey. 2025. (Direct URL not publicly indexed. Access via deloitte.com/nl or search “Deloitte Netherlands Gen AI ROI Survey 2025”.)

Accenture. Six Key Insights for C-Suite Executives to Maximize Return on Agentic AI. 2025. https://www.accenture.com/us-en/insights/technology/maximize-return-agentic-ai (Verify — Accenture restructures insights pages. Search accenture.com for “six insights agentic AI c-suite” if URL redirects.)

Accenture. The New Rules of Platform Strategy in the Age of Agentic AI. 2026. https://www.accenture.com/us-en/insights/technology/platform-strategy-agentic-ai (Verify — search accenture.com for “new rules platform strategy agentic AI” if URL redirects.)

OpenAI. The State of Enterprise AI: 2025 Report. December 2025. https://openai.com/enterprise/state-of-enterprise-ai-2025 (If URL redirects, search openai.com for “state of enterprise AI 2025” — OpenAI has moved several reports to gated download pages.)

Microsoft. Getting Started with Copilot. 2025. https://adoption.microsoft.com/en-us/copilot/

Microsoft. Copilot for Executives. 2025. https://adoption.microsoft.com/en-us/copilot/copilot-for-executives/

Microsoft. Becoming a Frontier Firm. 2025. https://www.microsoft.com/en-us/worklab/work-trend-index/2025-work-trend-index-annual-report (Microsoft Work Trend Index 2025 is the closest verified URL for frontier firm framing.)

World Economic Forum. Proof Over Promise: Insights on Real-World AI Adoption from 2025 MINDS Organizations. 2026. PDF: https://reports.weforum.org/docs/WEF_Proof_over_Promise_Insights_on_Real_World_AI_Adoption_from_2025_MINDS_Organizations_2026.pdf

KPMG. Boardroom View on Gen AI Adoption. 2025. https://kpmg.com/us/en/articles/2025/boardroom-view-generative-ai.html (Verify — KPMG report landing pages vary by region. Search kpmg.com for “boardroom view gen AI 2025” if URL redirects.)

Google Cloud. ROI of AI 2025. 2025. https://cloud.google.com/roi-of-ai (Google Cloud periodically updates this page in place rather than versioning it.)

Institute of Directors. AI Governance in the Boardroom. 2025. https://www.iod.com/resources/research/ai-governance-in-the-boardroom/ (Verify URL — IoD UK restructures resources periodically. Search iod.com for “AI governance boardroom” if URL redirects.)

Australian Institute of Company Directors. AI Use by Directors and Boards. 2025. https://aicd.com.au/research/governance-leadership-review/2025/ai-use-by-directors-and-boards.htm (Verify URL — AICD research pages require member login for full access in some cases.)

OECD. AI and the Global Productivity Divide: Fuel for the Fast or a Lift for the Laggards? 2025. https://www.oecd.org/en/publications/ai-and-the-global-productivity-divide_f8c45054-en.html


Supplementary Sources (added for open-source and cost-of-failure coverage)

European Union. Regulation (EU) 2024/1689: Artificial Intelligence Act, Article 99 — Penalties. 2024. https://artificialintelligenceact.eu/article/99/

IBM Security. Cost of a Data Breach Report 2025. 2025. https://www.ibm.com/reports/data-breach

Menlo Ventures. 2025: The State of Generative AI in the Enterprise. 2025. https://menlovc.com/perspective/2025-the-state-of-generative-ai-in-the-enterprise/

Stanford HAI. 2025 AI Index Report. 2025. https://hai.stanford.edu/ai-index/2025-ai-index-report


This article was synthesized by an AI governance and human-AI collaboration practitioner using the research collection curated by Alex Issakova. All data points attributed to specific organizations are drawn from publicly documented case studies in the source reports cited above. Conflicts of interest in source materials are noted in the text where applicable. The HAIA-RECCLIN governance framework, Checkpoint-Based Governance (CBG), GOPEL specification, and Factics methodology informed the analytical structure and practitioner commentary throughout. These frameworks represent one practitioner’s approach to the governance challenges the research documents; they are not presented as the only architecture capable of addressing them.

Share this:

  • Share on LinkedIn (Opens in new window) LinkedIn
  • Share on Facebook (Opens in new window) Facebook
  • Share on Mastodon (Opens in new window) Mastodon
  • Share on Reddit (Opens in new window) Reddit
  • Share on X (Opens in new window) X
  • Share on Bluesky (Opens in new window) Bluesky
  • Share on Pinterest (Opens in new window) Pinterest
  • Email a link to a friend (Opens in new window) Email

Like this:

Like Loading...

Filed Under: AI Artificial Intelligence, AI Thought Leadership, Thought Leadership, White Papers Tagged With: agentic AI, AI deployment, AI failure rates, AI Governance, AI pilots, AI research synthesis, AI ROI, AI safety, Basil Puglisi, boardroom AI governance, Checkpoint-Based Governance, enterprise AI, EU AI Act, Factics, FLI AI Safety Index, GOPEL, HAIA-RECCLIN, Human-AI Collaboration, McKinsey AI, OECD AI, open-source AI

Reader Interactions

Leave a Reply Cancel reply

You must be logged in to post a comment.

Primary Sidebar

Buy the eBook on Amazon

FREE WHITE PAPER, MULTI-AI

A comprehensive multi AI governance framework that establishes human authority, checkpoint oversight, measurable intelligence scoring, and operational guidance for responsible AI collaboration at scale.

SAVE 25% on Governing AI, get it Publisher Direct

Save 25% on Digital Factics X, Publisher Direct

Digital Factics X

For Small Business

Facebook Groups: Build a Local Community Following Without Advertising Spend

Turn Google Reviews Smarter to Win New Customers

Save Time with AI: Let It Write Your FAQ Page Draft

Let AI Handle Your Google Profile Updates

How to Send One Customer Email That Doesn’t Get Ignored

Keep Your Google Listing Safe from Sneaky Changes

#SMAC #SocialMediaWeek

Basil Social Media Week

Legacy Print:

Digital Factics: Twitter

Digital Ethos Holiday Networking

Basil Speaking for Digital Ethos
RSS Search

@BasilPuglisi Copyright 2008, Factics™ BasilPuglisi.com, Content & Strategy, Powered by Factics & AI,

%d