Surviving AI

The AI Wrapper Trap

Our first instinct was to bolt a chatbot onto 14 years of training content. Here is exactly what happened - and why every business reaches for the same dangerous solution first.

Mark Jones
Mark Jones · Collab365

Before you read what happened to us - consider what you would do right now. If someone told you tomorrow that your core content, your training library, your knowledge base, needed AI on top of it by end of quarter - what would you reach for? Almost certainly: a chatbot. Maybe Copilot. Maybe a retrieval layer sitting over your SharePoint. Something you can ship quickly. Something that looks like progress.

We reached for exactly the same thing. This chapter is about what we found when we got there.

When Collab365 realised the traditional video course format was dying, our first instinct was exactly the same as everyone else's.

"Let's just build an AI chatbot."

The plan was simple. We would take 15 years of Collab365 training videos, course transcripts, and blog posts. We would dump all of it into a vector database, bolt an AI wrapper on top, and call it a day.

It sounded brilliant. We thought it was going to save us thousands of hours.

There was one complication we had not anticipated. Many of our members had purchased lifetime access to this library. We had a legal and ethical commitment to honour that. So the legacy content had to come with us. It went into an archive: a searchable vault for existing members to access what they had paid for. That was the right thing to do.

But the archive preserved our obligation. It did not create a future.

A wrapper might help members search our old library, but it did absolutely nothing to fix our biggest threat: the velocity problem. Finding old data faster is useless if the software has already changed. We still had no high-speed way to create up-to-date content that solved real problems. If Microsoft shipped an update, we would still have to spend weeks manually recording videos just to feed the chatbot.

A smart search bar on top of a slow production line still leaves you permanently behind.

And when we actually tested the chatbot, we realised it couldn't even handle the legacy data properly. It walked us straight into a catastrophic trap.

The Frankenstein Phase & The App Sprawl Audit

To understand why the wrapper failed, you have to look at what we were actually wrapping. Over a decade, we had bolted together a sprawling Frankenstein's monster of platforms.

We had LearnDash for courses. WordPress for articles. WooCommerce for carts. Circle for community. We used ActiveCampaign for emails. Stripe for subscriptions.

If a member asked our new AI: "How do I fix the Power Automate throttling error my team hit yesterday?"

  • Their subscription state was in Stripe.
  • The specific error they were hitting was discussed by a user in Circle.
  • The structural solution was a video hidden in LearnDash.
  • The latest Microsoft patch note they needed was in an ActiveCampaign newsletter.

That single query touched data in at least four different systems. To answer it correctly, you'd need custom integrations bridging everything. We tried building them. We got things querying across platforms.

And then we stepped back and looked at what we had actually built.

It was a new layer of fragile AI infrastructure wrapped around the same disjointed data underneath, held together with API calls, token budgets, and a quiet prayer that none of those external platforms changed their code.

Architecture Note

What about the Model Context Protocol?

If you follow AI engineering, you might be wondering why we didn't just solve all of this by using MCP.

Model Context Protocol (MCP) is a new open standard that acts like a universal adapter plug. It allows an AI to hook directly into your external tools—like Stripe, Google Drive, or an old WordPress database—without you having to write messy custom code for every single connection.

It sounds like a magic bullet. For internal team tools and personal productivity bots, it absolutely is.

But for a commercial application serving thousands of customers, it becomes a heavy anchor. Every time a customer asks a question, the AI has to fetch huge amounts of raw data across the internet before it can even start thinking. Your system becomes slow, incredibly expensive, and entirely dependent on third-party tools you do not control.

The harsh reality: you are still wrapping an AI over a messy foundation you don't own. Moving your data into a single, lightning-fast architecture that you control will always beat passing massive data payloads blindly between external vendors.

The Hallucination Engine

Worse still, even when it could pull the data, the AI had no idea how to resolve conflicts across those systems.

We had a 2018 course on "Creating Flows". We had a 2021 update. And we had new material from 2024.

When we asked our shiny new AI bot how to build an approval workflow, it did exactly what it was programmed to do. It grabbed a button location from the 2018 transcript, a deprecated concept from 2021, and synthesised them into an incredibly authoritative-sounding answer.

An answer that was entirely, completely wrong for the 2026 interface.

This wasn't fixing our education problem. It was automating confusion at machine speed. We were building a Hallucination Engine.

And we quickly discovered this wasn't just a failure of our little prototype. It was the defining failure of enterprise AI worldwide.

Research from the MIT NANDA initiative established that 95% of generative AI pilots fail to reach production with any measurable business impact.MIT Sloan

Gartner warned that at least 30% of all generative AI projects would be fully abandoned post-pilot by the end of 2025. The average sunk cost per abandoned initiative: $7.2 million.Gartner It is the GIGO principle: Garbage In, Garbage Out. AI does not rectify a messy data foundation. It regurgitates it.

The Air Canada Verdict

Air Canada thought a hallucinating chatbot was just an edge case. Their tribunal ruling changed that assessment permanently.

In November 2022, a customer used Air Canada's AI chatbot to ask about bereavement fares. The bot confidently told him he could book full price and claim a retroactive refund. He followed the instructions exactly. Air Canada later denied the refund, citing their actual written policy.

Air Canada tried to argue the chatbot was a "separate legal entity" responsible for its own actions.

The Tribunal rejected this entirely. The ruling (2024 BCCRT 149) established that companies are strictly and fully liable for the damages caused by their hallucinating AI tools.American Bar Association

The "bolt an AI wrapper onto old, disjointed systems and hope for the best" strategy is a ticking time bomb.

95%

The percentage of enterprise generative AI pilots that fail to reach production with any measurable business impact. The bolt-on wrapper approach simply does not work.

The Hard Pivot

We were not alone in this. Law firms pointing Copilot at decades of unstructured case files. Insurance companies letting AI summarise policy documents that contradict each other. Financial advisers deploying chatbots grounded in out-of-date regulatory guidance. The wrapper trap is not a training company problem. It is the defining failure pattern of the first wave of enterprise AI adoption.

We refused to ship it. If we were going to build something worth subscribing to, it could not be a wrapper that hallucinated its way through members' real problems.

To truly use AI Agents across your business, you don't need better connectors between your fractured platforms. You need one platform, one codebase, and one database.

When everything lives in one architecture, AI can reach across every part of the business. Member data, content, subscriptions, community discussions: all of it queryable, all of it grounded in truth. Trying to achieve that across four separate platforms you do not own is not just slower. It is structurally impossible.

So Collab365 did not just build a new UI. We did the brutal work of extracting our entire ecosystem into a single, clean, AI-readable architecture. And we completely redefined what the system would produce.

The Scope Had Changed

By the time we had faced down the data quality problem and the hallucination problem, we had stumbled into a third problem that was arguably the biggest of all.

The industry we had been training people to navigate no longer existed in the shape it had been.

We had been a Microsoft 365 training company. That felt like a safe, defined scope: one ecosystem, one audience, one established partner. But AI did not arrive as a Microsoft product. It arrived everywhere at once. OpenAI. Anthropic. Google. Meta. Mistral. New models and new capabilities were landing every few weeks from vendors we had never trained anyone on.

Microsoft did not even have an established large language model of their own when ChatGPT shipped. They had to scramble to integrate OpenAI into what became Copilot. The defined, stable Microsoft ecosystem that had anchored our business for fourteen years was now just one player in a much larger and faster-moving field.

Our members stopped asking how to use Power Apps. They started asking how to evaluate which AI model was right for their workflow. How to build automation that worked across tools from different vendors. How to deploy agents in platforms that had not existed six months earlier.

Software-specific, vendor-specific training had stopped cutting it. The questions had become vendor-agnostic, and so had the answers.

We needed to stop teaching people how to use specific software and start teaching them the underlying strategies and techniques of AI and automation that work regardless of which vendor they happen to be using this month.

Proving the Theory First

We knew the wrapper was a dead end. We knew we had to build an entirely new, single architecture that native AI could read and trust. But burning down fourteen years of legacy platforms to build a clean engine from scratch was a massive, risky pivot.

Before we committed to that pain, we had to know if the destination was actually worth it. Could humans and AI actually work together to produce world-class, vendor-agnostic problem-solving content at an industrial scale? We had to prove the theory.

Before we built anything new, we ran an experiment. We forced ourselves to use AI to generate new content completely manually. It nearly broke us.