

AI leadership increasingly requires a hard call: deciding when an AI pilot is learning fast enough to justify continued investment—or when it is time to stop.
In CIO media coverage about abandoning AI proofs of concept, Andreas Welsch, an independent AI advisor, emphasizes that the decision becomes clearer when leaders define success metrics early and revisit them at predetermined checkpoints.
This article is written for CIOs, CTOs, and other executives overseeing AI adoption, AI governance, and AI strategy, with a focus on avoiding prolonged “pilot purgatory” while preserving room for informed experimentation.
Original source: When is the right time to dump an AI project?
Why this media coverage matters
The CIO article frames a common executive dilemma: cancel an underperforming AI initiative too early and potential future value may be missed; keep it running too long and costs, time, and opportunity loss compound.
The audience is enterprise leadership responsible for translating AI ambition into operating results. The coverage is relevant to AI leadership because it connects operational governance (metrics, checkpoints, stakeholder alignment) to practical decision-making in AI adoption—especially when projects fail to meet business objectives.
Executive Summary
- Define success metrics beyond ROI before launching an AI pilot.
- Set timelines and checkpoints to evaluate progress and stop drift.
- Tie each AI project to a specific business problem and stakeholder need.
- Measure outcomes carefully, including hard-to-quantify productivity gains.
- Act early to avoid sunk-cost dynamics and long “pilot purgatory.”
Key Takeaways
- Welsch warns that projects often continue because teams “hope” a breakthrough is near.
- Welsch recommends establishing proper goals and KPIs early, then checking progress on schedule.
- Welsch stresses that AI efforts should start with a defined business problem to solve.
- Welsch advises close collaboration with business stakeholders at the beginning of the initiative.
- When KPIs are missed, leadership can decide to stop, extend, or recalibrate the project scope.
- Small KPI shortfalls (for example, 7% vs. 10%) may still warrant continued investment.
What is AI leadership?
AI leadership is the executive practice of aligning AI initiatives to business objectives, setting clear success metrics, and creating decision cadence so AI projects can be scaled, corrected, paused, or stopped. In the CIO coverage, Andreas Welsch highlights that unclear goals can keep teams moving forward on hope rather than evidence. AI leadership also includes partnering closely with business stakeholders to ensure the work solves real pain points and can be adopted, rather than operating as disconnected experimentation.
AI leadership begins before the pilot: define success beyond ROI
One of the earliest governance moves is deciding how success will be evaluated. Andreas Welsch points to a common failure pattern: teams pursue an AI initiative without “proper goals in place,” then struggle to justify stopping because they keep expecting a breakthrough.
In the CIO coverage, examples of KPIs that can anchor an AI pilot include increasing customer satisfaction by 10%, reducing the time to fill out a request for proposal (RFP) by 30%, or spending four fewer hours per month paying invoices. These are operational targets leaders can revisit at set checkpoints.
Key Insight: Andreas Welsch argues that the “right time” to stop becomes visible when leadership sets KPIs and a timeline upfront. Without predetermined goals and checkpoints, teams often continue simply because they expect a breakthrough “right around the corner,” even when progress is unclear.
Checkpoint discipline reduces emotion-driven decisions
Welsch recommends predetermined checkpoints where IT and business teams assess progress toward agreed metrics. If the initiative is not meeting targets, leadership can decide whether to stop it, adjust it, or give it more time. The key is that the evaluation is planned—not improvised after months of effort.
The CIO article notes that AI pilots can become expensive, and that a structured checkpoint approach helps avoid prolonged investment with limited value.
Connect AI projects to business needs to avoid “AI for AI’s sake”
AI initiatives often fail basic alignment: they are launched to “do something with the technology” rather than to address a specific pain point. Welsch’s guidance in the CIO coverage is direct: leaders should identify the business problem being solved and work closely with stakeholders from the start.
When the business need is explicit, adoption is easier to evaluate and outcomes are simpler to measure. When it is not, teams may build capabilities that do not fit real workflows, creating a project that is difficult to scale—or even to fairly assess.
Key Insight: Welsch’s stakeholder-alignment point is a governance lever, not just a planning tip. When business stakeholders and IT agree on the problem statement early, the project earns a clear “license to operate,” and leadership can judge continuation based on business outcomes rather than technical progress alone.
Operational examples that leadership can validate
The CIO coverage offers concrete KPI examples executives can validate with stakeholders: customer satisfaction improvement, RFP cycle-time reduction, or fewer hours spent on invoice payments. These targets also help determine whether the project should continue even if the result is slightly below the initial goal.
Make measurement credible, especially when benefits are “less tangible”
Even with KPIs, measurement can be hard. The CIO coverage highlights that some benefits are evasive—such as estimating time saved when employees use an AI copilot to draft an email. If measurement misses these effects, leadership may mistakenly conclude that value is absent.
That risk reinforces the need for AI governance practices that define how success will be measured—not only what will be measured. When leadership decides to stop an AI pilot, the decision should rest on evidence that reflects the initiative’s real outcomes.
Use “directional” evaluation when precision is unrealistic
The CIO article suggests that ROI can emerge from multiple smaller effects, not just a single, easily audited number. Executives can still set measurable targets, but should recognize that some assessments may be directional, especially early in adoption.
Limit the damage: time-box experimentation to avoid sunk-cost dynamics
Cost and time escalation are central to the “pull the plug” question. The CIO coverage cites Gartner estimates that a retrieval-augmented generation (RAG) AI document search project can cost up to $1 million to deploy, with recurring per-user costs of up to $11,000 a year. It also notes that a medical, insurance, or financial large language model (LLM) built from scratch can cost up to $20 million.
Given those stakes, the coverage includes the view that long pilots can become dangerous when they run six to nine months: leadership may fall into sunk-cost thinking and try to force a project to work after substantial effort and spend.
Key Insight: High AI project costs make governance cadence a financial control. The CIO coverage shows why leadership should time-box evaluation: when pilots run for months without decisive checkpoints, sunk-cost pressures rise, and decisions shift from evidence-based governance to reputation and effort protection.
Failure rates are improving, but abandonment remains material
The article cites Gartner: in 2022, nearly half of AI pilots failed to reach production, while Gartner expects about 30% of generative AI projects will be abandoned after proof of concept by the end of 2025. Even with improvement, widespread experimentation means the aggregate cost of abandoned projects can be significant.
The coverage also cites an EY survey published in July, reporting that 95% of senior executives said their organizations were investing in AI—raising the importance of decision discipline across many simultaneous initiatives.
Dump, pause, or refocus: three distinct executive decisions
The CIO coverage includes three “stop-paths” leaders can consider when an AI pilot fails to meet expectations: dumping the effort, pausing it, or refocusing it with clearer objectives.
Welsch’s contribution centers on preventing the need for late-stage rescues by setting goals early. However, the coverage also notes that it can sometimes be early enough to refocus a project by clarifying the end goal when aims are open-ended and the work lacks focus.
Use KPI shortfalls to guide a rational next step
The CIO article gives a pragmatic example: if a customer service bot improves satisfaction by 7% instead of 10%, leadership may still consider continued investment. The decision hinges on whether the shortfall is acceptable relative to business value, cost, and the feasibility of closing the gap.
AI leadership in practice: building a repeatable “stop” mechanism
“When to stop” is not only a project-management question; it is a leadership operating model. In the CIO coverage, Welsch’s guidance implies a repeatable mechanism: start with a business problem, define KPIs beyond ROI, set checkpoints, and evaluate progress against outcomes rather than hope.
That mechanism supports adoption decisions across a portfolio of AI experiments—especially as AI investment becomes widespread and pilot failure remains common.
Leadership Implications
- Codify success metrics early: Define KPIs beyond ROI and document how they will be measured at launch.
- Set governance cadence: Establish predetermined checkpoints to evaluate progress and make continuation decisions.
- Anchor to business pain points: Start every AI initiative with a specific problem statement agreed with stakeholders.
- Design for adoption evidence: Ensure measurement captures both operational metrics and harder-to-quantify productivity gains.
- Avoid sunk-cost traps: Time-box pilots so stopping is a normal governance act, not a late-stage crisis.
Why this matters now
The CIO coverage positions AI as both promising and costly, with pilots that can become money pits if they produce no value. With many organizations investing in AI, leaders need consistent AI governance and AI strategy practices to manage experimentation at scale.
Andreas Welsch’s emphasis on clear goals, metrics, and stakeholder alignment reinforces a broader AI leadership mandate: decisions should be made on business outcomes and evidence, not momentum. That mindset supports responsible AI adoption and helps organizations avoid extended “pilot purgatory.”
Conclusion
AI leadership is tested when projects underperform. The CIO coverage shows that the most defensible “stop” decisions come from upfront clarity: define success metrics beyond ROI, tie the initiative to a real business need, and evaluate progress at predetermined checkpoints.
With AI pilots carrying significant costs and failure still common, leaders who normalize evidence-based continuation decisions can protect budgets, improve adoption outcomes, and build a healthier pipeline of AI initiatives that earn the right to scale.
About the Author
FAQ
When is the right time to dump an AI project?
The right time is when predetermined checkpoints show the pilot is not progressing toward defined success metrics tied to business objectives. AI leadership requires making this decision based on evidence, not hope that a breakthrough is imminent. Clear KPIs make the call defensible.
What KPIs should executives set for an AI pilot?
Executives can set operational KPIs such as improving customer satisfaction by 10%, reducing RFP completion time by 30%, or saving four hours per month on invoice payments. AI leadership uses these metrics beyond ROI to judge progress at scheduled checkpoints and decide whether to continue.
Why do AI pilots get stuck in “pilot purgatory”?
AI pilots get stuck when teams lack proper goals, measurement plans, and deadlines, so the project continues on optimism rather than outcomes. AI leadership prevents drift by defining success early and revisiting it at predetermined checkpoints. Without this, stopping feels arbitrary and late.
How should AI projects be tied to business needs?
AI projects should begin with a clearly defined business problem and close collaboration with business stakeholders from the start. AI leadership treats alignment as a requirement for adoption, not a nice-to-have. When the problem statement is explicit, value and progress are easier to assess.
Should AI success be measured only by ROI?
No—AI leadership should define metrics beyond ROI because some benefits are less tangible and hard to capture in a single financial number. Measurement should include operational outcomes and productivity effects. Poor measurement can cause leaders to abandon useful initiatives prematurely due to incomplete value signals.
How can leaders avoid sunk-cost fallacy in AI initiatives?
Leaders can avoid sunk-cost fallacy by time-boxing pilots and scheduling go/no-go checkpoints early, before months of spend and effort accumulate. AI leadership normalizes stopping as governance, not failure. Longer pilots can make teams feel pressured to “force” success after heavy investment.
How expensive can AI projects become?
AI projects can be costly: a RAG AI document search deployment can reach $1 million, with recurring per-user costs up to $11,000 annually, and a domain LLM built from scratch can cost up to $20 million. AI leadership treats cost visibility as essential to stop decisions.
Is a near-miss on KPIs a reason to stop an AI project?
Not necessarily; AI leadership can treat a KPI near-miss as a decision point to adjust scope, extend time, or continue investment based on business value. The CIO coverage gives an example: a bot improving satisfaction by 7% versus 10% may still justify continuation depending on context.
What is the role of checkpoints in AI governance?
Checkpoints create a predetermined cadence for evaluating progress against success metrics and business objectives. AI leadership uses checkpoints to make continuation decisions routine, evidence-based, and cross-functional. Without them, teams often delay decisions until costs and organizational commitments make stopping difficult.

