Surviving AI

The Zero-Server Ecosystem

Why we threw away Microsoft Azure, built an orchestration layer to avoid OpenAI lock-in, and now run the entire engine with zero servers.

Mark Jones
Mark Jones · Collab365

An engine that constantly discovers problems, verifies logic, and synthesizes courses sounds incredibly expensive to run. In a traditional corporate environment, it is. Running multiple persistent AI agents traditionally requires a dedicated DevOps team, massive cloud instances, and continuous maintenance.

We are a team of two people. We have exactly zero servers.

The architecture that powers Collab365 Spaces relies on two aggressive decisions that go against standard enterprise thinking. The first was our hosting provider. The second was our AI model strategy.

Why We Walked Away From Microsoft Azure

Our entire business was built around teaching Microsoft 365. Naturally, when we first architected the intelligence engine, we prototyped the entire thing in Microsoft Azure.

It took way too long.

Azure is an incredibly robust, enterprise-grade platform. But for a nimble team trying to execute at extreme velocity, Azure presented overwhelming decision fatigue. There were too many ways to crack a single problem. The configuration was heavy. The deployments were slow. The architecture forced us to spend more time managing infrastructure than writing business logic.

We scrapped the prototype and moved the entire ecosystem to Cloudflare.

The difference in velocity was staggering. Cloudflare allowed us to deploy code globally to the edge (servers located physically close to your users around the world) in seconds, bypassing Kubernetes clusters and virtual machines entirely. But beyond deployment speed, it gave us an entire architectural stack built specifically to handle AI:

  • Cloudflare AI Gateway: We route every single AI request through this gateway. It gives us unified instrumentation, rate limiting, and exact cost-tracking for every prompt across multiple payment providers in one simple dashboard.
  • Cloudflare Workflows: This is the backbone of the architecture. We have dozens of complex background workflows—a course workflow, a recipe workflow, an avatar generator. Cloudflare Workflows manages the state between every step, logs the outputs, reports on failures, and operates for fractions of a penny.
  • Cloudflare Vector Search: This powers our pristine Knowledge Base. It allows us to easily ingest deep research and instantly deploy semantic RAG search across the entire platform.
  • Hosted Edge Models: For simple utility tasks and data structuring, we use Cloudflare's own locally hosted edge models. They are incredibly fast and essentially free.

Vendor Agility: Never Marry One AI

The second major decision was about the AI models themselves. Most companies lock themselves into a single vendor—usually OpenAI—by hardcoding API calls straight into their application.

This is a fatal architectural mistake. If your provider degrades their model or raises prices, your entire business is held hostage.

To survive, we built an Orchestration Layer behind the AI Gateway. Nowhere in the Collab365 codebase is an AI model ID hardcoded. Instead, tasks are assigned a "Role":

  • For complex Deep Reasoning, we route to Google Gemini.
  • For heavy Extraction and Structuring, we route to Anthropic's Claude.
  • For fast Web Searches and Synthesis, we route to xAI's Grok. It is nearly five times cheaper for these specific calls and blindingly fast.xAI API Pricing

When a new model drops that is 10% smarter or 50% cheaper, we change one string in the orchestrator. The entire platform upgrades globally in under two minutes without pushing a single line of code. We preserve total vendor agility.

The Zero-Server Result

Because we rejected the heavy infrastructure of Azure and the vendor lock-in of single AI providers, we have a system that is infinitely scalable, completely resilient, and incredibly cheap to run.

You do not need an army of engineers to transform your business for the AI era. You need clarity of architecture. You need to stop building wrappers and start building ecosystems. And you need to make the hard decisions now, before your legacy debt mathematically eliminates you from the market.

You have seen the architecture, the mistakes, and the strategy. Now it is time to turn the mirror on your own organisation. Are you building an intelligent asset, or are you just wrapping a bloated, broken legacy machine?

2 Minutes

The exact time it takes to switch out our core reasoning AI model across the entire Collab365 architecture if a competitor drops something cheaper and faster.