
Everyone is talking about building AI applications. Pick an LLM, send it a prompt, get a response. Ship it. Iterate. Move fast and break things.
This works wonderfully for hackathons, demos and weekend projects. But it does not work for your real-world customers.
The statistics I saw in the last year are different than an idyllic world: around 60/70% of enterprise AI projects fail. Only around 5% of AI pilot programs achieve rapid revenue acceleration. The average organization scrapped around 40% of AI proof-of-concepts before they reached production.
Everyone talks about and celebrates “vibe coding”: building AI applications with remarkable speed using tools like GitHub Copilot, Claude, Codex and more. Anyone can now build something that feels intelligent. Everyone talks about rules for vibe coding, new tools, best practices, N coding agents to increase the vibe coding experience. All good and exciting (I’m a fan of this too), but this is not what customers want (and we’re honestly quite tired of always talk about those stuffs nowadays ).
In the enterprise, AI is categorically different.
When a customer asks you to build an AI agent for their business, they are not asking for a toy. They are asking for a system that:
Vibe coding ignores all of this. It assumes a stable environment, predictable behavior, and the luxury of rebuilding when something breaks. Enterprise AI assumes none of these things.
Customers do not want vibe coding. They want:
None of these aspects generates the viral engagement of a social post about the latest vibe coding features and capabilities. But all of it is what actually matters to customers.
The gap between what the AI ecosystem talks about and what customers need is widening. The ecosystem has become a feedback loop of tools for tools, frameworks for frameworks, optimization of the demo. Meanwhile, customers are sitting with real problems that require something the ecosystem largely ignores: reliable, integrated, governance-aware enterprise AI.
I wrote this provocative sentence some weeks ago on LinkedIn and it generates quite a lot of discussions.
The simplest mental model for AI is a function: input goes in, the LLM processes it, output comes back. This is how most demos work. This is not how business systems work.
A real-world AI agent in an enterprise context typically needs to:
The LLM is one piece of a much larger puzzle. It is a powerful reasoning engine, but it is not the system itself. The system is the orchestration around it.
The big mistake I often see is treating the LLM as the solution. In reality, the LLM is a component. The solution is the architecture that puts that component to work reliably within constraints, policies, and integration points that matter to the customer. Building an enterprise AI system requires orchestrating multiple services.
Consider a realistic scenario: an AI assistant that helps a customer service team handle complex inquiries related to orders, inventory, and shipping.
The workflow might look like:
This realistic flow requires integration with:
Each integration has its own failure modes, latency characteristics, quota limits, authentication requirements, and data format expectations. Some are synchronous, others support webhooks. Some are rate-limited, others have per-request costs. Some require retry logic, others don’t tolerate duplicate calls. The system must handle all of them gracefully, in parallel, and without losing state.
In enterprise AI systems, the orchestration layer (the logic that coordinates services, handles errors, manages state, manages rate limits, reconciles conflicting data, and enforces business rules) often contains more value and complexity than the LLM integration itself. It is the core of your architecture, so treat it as such.
Another problem I often see is the following: an AI workflow can take minutes, hours, or even days to complete, far too long for a synchronous HTTP request. The client cannot hold a connection open. The process must survive system restarts, network failures, and manual intervention.
Enterprise-grade AI systems should use asynchronous patterns:
This pattern requires infrastructure that can handle persistence, retries, and scheduling. Cloud services are almost essential here, like for example:
These are not optional luxuries. They are foundational requirements for systems that need to survive failure and handle workloads that exceed the limitations of synchronous request-response patterns.
Designing an agent architecture for a real-world business process is complex and requires different choices.
Agents need access to tools—functions they can call to interact with the business system. These might include:
The LLM decides which tools to use and when. The orchestration layer ensures the tools behave correctly and handle errors. This division of responsibility is critical: the LLM reasons, but it does not execute directly. Execution is controlled and validated.
Agents need guardrails (constraints that prevent them from making dangerous decisions). Examples:
These rules are not expressed in the prompt. They are enforced by the orchestration layer. The agent can request an action, but execution only proceeds if the action passes validation.
Complex tasks require agents to reason across multiple steps, potentially asking for clarification or human input. An agent building a sales proposal might:
This requires state management, context preservation, and graceful handling of human interaction points. As you can see, it is far more complex than a single LLM call.
Business Central is a powerful ERP system, but it is not designed as an AI-native platform for everything. Integrating AI agents with Business Central introduces specific challenges that often requires to use also external tools and planning the architecture of the solution accordingly.
Some organizations look at Business Central’s built-in AI features or Copilot Studio and assume those tools are sufficient for building enterprise AI agents. They are not. Not for meaningful business transformation.
Built-in AI features in Business Central are designed for simple, isolated tasks. They work within the boundaries of Business Central’s data model and cannot easily orchestrate workflows that span multiple systems.
Copilot Studio is more flexible, it allows you to build conversational agents and wire them to connectors. But at its core, it is a low-code platform optimized for rapid prototyping and demo-ready solutions. It is not built for:
Copilot Studio can be a component of your solution (a UI layer, perhaps, or a proof-of-concept starting point). But treating it as your AI architecture is like building a skyscraper with a prefab shed. It works for the walls, but the structural system needs to be engineered separately.
Here is the hard truth (at least in my opinion) about why so many customer’s AI projects fall short of expectations: many organizations take an existing workflow they have in place and put some AI on top of it. They automate one step, maybe two. But the fundamental processes, roles, dependencies, controls, bottlenecks and overall work logic remain intact.
Real transformation does not occur when we ask AI to do an existing task better. It occurs when we ask whether that task should exist in that form at all. It occurs when we stop thinking about automating individual activities and start thinking about redesigning workflows.
This is IMO why so many initiatives fail or stall at modest results: it is far easier to buy software than to fundamentally restructure culture, accountability, organizational interfaces, and processes that have been refined over years.
The wrong question is: Where can we put AI?
The right question is: Which parts of our work need to be fundamentally redesigned so that AI actually matters?
This is not a semantic difference. It is a strategic difference. In the first case, you are adding a technology layer. In the second, you are redesigning an operating system. Customers who ask the right question are the ones who see transformational results. Customers who ask the first question tend to see only modest and localized improvements.
This is a topic where I personally fight with many (internally in my company and externally with partners). Everyone now is an expert on AI just because there are some agents in Business Central or there’s Copilot Studio or there’s the vibe coding option to create something.
But building business AI agents that deliver real value requires developer-oriented skills and technical depth. This is not low-code work. This is engineering.
No-code and low-code platforms lower the barrier to entry for simple, isolated AI use cases. They do not eliminate the need for serious engineering when the stakes are high, the scope is broad, and the complexity is real.
I think that if you are a systems integrator, partner, or consultant proposing AI solutions to customers, you need capabilities that go far beyond knowing how to call an LLM.
First of all, you need to understand how distributed systems work: service orchestration, async patterns, state management, failure modes, and scalability constraints. This is not a “nice to have”, it is foundational. Vibe coding does not scale to enterprise workloads.
Then, you need to help customers ask the right questions about their workflows. This is not a technical work, but it is more a strategic and organizational one. You are helping them see where AI can enable fundamentally different ways of working, not just faster execution of existing processes. This requires:
Building enterprise AI systems requires solid cloud infrastructure and services knowledge: Azure Functions, Logic Apps, Service Bus, SQL Database, application monitoring, and logging. You need to understand:
Yes, you also need to understand LLMs: their strengths, limitations, failure modes, and how to prompt them effectively. But this knowledge is only valuable in context. You need to know:
This is harder than reading the latest LLM paper. It is about applying LLM knowledge to real business constraints.
Developing these skills takes time and real projects. It is not something you can absorb from online courses alone. You need to build real systems, fail at them, learn from failures, and refine your understanding. This is the barrier to entry for credible AI partners (and why so few exist who can execute at this level).
Oh, I want also to mention here a last topic that I think it’s really important: do honest conversation about AI limitations with your customers. Tell customers what AI cannot do. Can you protect your customer’s budget from an overhyped, scope-creeping project that promises 90% automation when 40% is realistic? Can you push back on unrealistic timelines? Can you clearly articulate the risks and effort involved?
Partners who do this earn customer trust and deliver sustainable results. Partners who don’t eventually deliver disappointing projects and damaged relationships.
Customer AI projects fail not because AI is overhyped or because the technology doesn’t work (it does). They fail because the gap between vibe coding and enterprise AI is enormous and most organizations, partners, and customers underestimate it.
They fail because adding a conversational layer on top of existing processes is not the same as transforming how work gets done.
They fail because building a reliable, scalable, integrated AI agent requires architecture, domain knowledge, process redesign, cloud infrastructure, and honest problem-solving that goes far beyond knowing how to send a prompt to an LLM or just using tonly he standard AI features that you have inside a platform like your favourite ERP .
If you are serious about AI for your customers, ask the right questions. Not “where can we put AI?” but “what should we fundamentally reimagine?” Not “can we build this quickly?” but “can we build this sustainably?” Not “how do we add a chatbot?” but “how do we redesign this workflow so that human intelligence and machine intelligence complement each other?“
The organizations that ask these questions and commit to the work will see transformational results. The others will see expensive pilots that deliver modest improvements and slowly fade away.
The choice is yours. So is the responsibility for making it honestly.
Original Post https://demiliani.com/2026/05/04/why-customers-ai-project-fail-the-gap-between-hype-and-reality/