Skip to content
Back to Blog

Your AI Agents Don’t Need to Be Smart. They Need to Be Cheap.

AgencyBoxx Team
Your AI Agents Don’t Need to Be Smart. They Need to Be Cheap.

There is a moment in every agency owner's AI journey where the credit card statement arrives and the math stops making sense.

You built the agents. They work. The email triage is saving you an hour a day. The time tracking enforcement is running without complaints. The prospecting pipeline is filling up. Everything is humming. Then you check the bill and realize you spent more on AI tokens last month than you would have spent on the junior employee the agents were supposed to replace.

We hit that wall early. Our first architecture burned through $150 in four hours. Even after we restructured to a script first approach, the AI portion of our system was still more expensive than it needed to be, because we were using cloud models for tasks that did not require cloud grade intelligence.

The fix was not smarter models. It was dumber ones. Running locally. For free.

The Problem with "Smart by Default"

When most people start building AI agents, they reach for the best model they have access to. Claude Opus. GPT 4. Gemini Pro. It makes sense intuitively: you want quality output, so you use the highest quality model.

But quality is not binary. It is a spectrum, and most agent tasks live at the low end of that spectrum.

Consider what a time tracking compliance agent actually does with AI (if it uses AI at all). It checks whether a time entry description meets a minimum quality threshold. That is a classification task. "Does this description contain meaningful information about the work performed, or is it blank/generic?" A $20 per million token model will give you the same answer as a model running for free on your own hardware. The task does not require world knowledge, complex reasoning, or nuanced language generation. It requires a yes or no judgment on a short string of text.

Now multiply that by every low complexity task across your entire agent roster: spam classification, email categorization, document summarization, data cleaning, format validation, sentiment detection on short text, keyword extraction, and embedding generation. These tasks run hundreds or thousands of times a day. If each one hits a cloud API, the tokens add up fast. Not because any single call is expensive, but because the volume is relentless.

The solution is to stop treating every AI task like it requires the smartest model in the room.

Local Models Changed the Math

We run Ollama on the same Mac Studio that hosts our entire agent infrastructure. Ollama is an open source tool that lets you download and run AI models locally. No API calls. No token costs. No data leaving the building.

The models we use for local processing:

Qwen 2.5 (and now 3.5): General purpose reasoning at a surprisingly high level. When Qwen 3.5 dropped, we tested it against tasks we had been sending to mid tier cloud models and found comparable quality on summarization, classification, and basic analysis. The 4 billion parameter version runs comfortably on modest hardware. The 9 billion version pushes the limits of our 36GB of RAM but handles more complex tasks.

nomic embed text: Embedding generation for our entire knowledge base. Every document chunk, every email, every meeting transcript gets embedded locally. This is the foundation of our RAG (retrieval augmented generation) system, and it runs at zero marginal cost regardless of volume. When you have 33,700+ indexed chunks and you are continuously adding more, the savings from local embeddings versus cloud embedding APIs compound fast.

Llama 3.1: An alternative general purpose model we keep available for tasks where a second opinion from a different model architecture is useful, or when Qwen is busy processing a batch.

The total cost of running these models: zero. They run on hardware we already own. The Mac Studio was a one time $2,000 CAD investment. It handles the local AI models alongside everything else: 50+ services, 20 agents, three OpenClaw instances, the full RAG database. No monthly compute bills. No scaling costs. No vendor lock in.

Why Baked In AI Falls Short

Every SaaS tool in your stack is adding AI features. ClickUp has AI. HubSpot has Breeze. Fireflies has AI summaries. Notion has AI. On paper, this should mean you do not need your own models. The tools you already pay for are giving you AI for free (or for a modest add on fee).

In practice, baked in AI has three problems that make it unreliable for serious agent operations:

You cannot choose the model. When HubSpot's Breeze agent tries to build something and gets it wrong, you have no recourse. You cannot swap in a better model. You cannot fine tune the prompt. You cannot feed it your own knowledge base for context. You get whatever model the vendor chose, with whatever prompt engineering they did, and you hope it works. Sometimes it does. Often it does not. We have tried multiple times to have HubSpot's built in AI perform complex tasks. The success rate did not inspire confidence.

You cannot control the context. Baked in AI only sees what the tool gives it. Fireflies AI can summarize a call, but it does not know your client history, your project budget status, or the email thread that preceded the meeting. ClickUp AI can help with task descriptions, but it does not know your SOPs or your client's communication preferences. The intelligence is siloed inside each tool, which means it can never synthesize across your operations the way a purpose built agent with access to your full knowledge base can.

You cannot control the cost. Some tools charge per AI query. Some bundle it into higher tier plans. Some throttle usage. You have no visibility into what model is running, how many tokens each query consumes, or how to optimize. When you control your own model stack, you see exactly what every operation costs and you can route tasks to the cheapest model that produces acceptable quality.

This is not a knock on any specific vendor. Building AI into a platform that serves millions of users is genuinely hard, and generalized AI features will always be less effective than purpose built agents with controlled context. But it means that the "just use the built in AI" approach tops out quickly for agencies with serious operational needs.

The Tiered Stack in Practice

Our production system routes every AI task to one of four tiers based on what the task actually requires:

Tier 1: Local models (zero cost). Embedding generation, document chunking, text summarization, data cleaning, basic classification, format validation. These tasks run thousands of times a day and would cost hundreds of dollars a month on cloud APIs. Running them locally costs nothing after the initial hardware investment.

Tier 2: Cheap cloud models ($0.30 per million input tokens). Gemini Flash for standard email classification, template driven content fills, and high volume tasks that need slightly more intelligence than a local model but do not justify a premium API call. This tier handles the middle ground: tasks where a local model occasionally gets it wrong but a premium model would be overkill.

Tier 3: Mid tier cloud models ($1.25 per million input tokens). Gemini Pro for email drafts, meeting briefings, agent coordination, and tasks that require genuine reasoning. This is the workhorse for most AI powered operations that are not directly client facing.

Tier 4: Premium cloud models ($2.00+ per million input tokens). Reserved exclusively for high stakes output: client facing email drafts, complex diagnostic analysis, and anything where getting it wrong has real consequences. Used only after a cheaper model has already compressed the context, so the premium model sees a focused brief instead of raw input.

The routing is not dynamic or AI driven. It is hardcoded per task type. Time tracking compliance: local. Spam classification: local. Email embedding: local. Meeting transcript summarization: Tier 2. Draft reply generation: Tier 3. Final client facing polish: Tier 4. Every task has a predetermined tier, and we only move a task to a higher tier if the lower tier demonstrably fails at acceptable quality.

This is the same logic you would use staffing a team. You do not have your senior strategist formatting spreadsheets. You do not have your junior coordinator writing client proposals. Match the cost of the resource to the complexity of the work.

The Knowledge Base Is Where Local Models Pay for Themselves Fastest

If there is one place where local models deliver disproportionate value, it is the knowledge base.

We scraped and indexed HubSpot's knowledge base, API documentation, and community forums into 30,000+ local documents. Each document gets chunked, cleaned (strip the headers, footers, navigation HTML, and get down to the useful content), and embedded. The cleaning step uses a local model to extract the signal from the noise: what is the problem, what is the solution, and what context matters. The embedding step converts each chunk into a vector for semantic search.

If we did this on cloud APIs, the embedding cost alone for 33,700+ chunks (and growing) would be a recurring expense every time we re index or add new content. Running it locally means we can re index the entire knowledge base on a whim at zero incremental cost. We can experiment with different chunking strategies, different embedding models, and different retrieval approaches without worrying about the bill.

When an agent queries the knowledge base, the retrieval is entirely local. The query gets embedded locally, the semantic search runs against the local vector database, and the matching chunks are returned. No tokens consumed. No API call. The only time a cloud model gets involved is if the agent needs to synthesize the retrieved chunks into a coherent answer for a human, and even then, the context has already been narrowed down to just the relevant chunks instead of the full 30,000+ document corpus.

This is the compounding advantage of local infrastructure. Every document you add to the knowledge base makes the system smarter. None of those additions cost anything to process. Over 12 months, the gap between "everything on cloud APIs" and "local first with surgical cloud usage" widens from hundreds of dollars to thousands.

What About Quality?

The obvious objection: do local models produce worse output?

For the tasks we route to them, no. Not in any way that matters operationally.

A local embedding model produces vectors that are functionally equivalent to cloud embedding APIs for retrieval purposes. A local model summarizing a meeting transcript to extract action items produces output that is 95% as good as a premium cloud model on the same task, at zero cost. A local model classifying an email as "spam," "newsletter," "client request," or "internal update" gets it right at the same rate as an expensive model because the task does not require the extra intelligence.

Where local models fall short is on tasks that require broad world knowledge, complex multi step reasoning, nuanced tone matching, or creative generation. Those tasks go to cloud models. That is the entire point of tiering: you are not choosing between quality and cost. You are choosing both, on a per task basis.

The quality concern we hear most often is: "But what if the cheap model gets it wrong and nobody notices?" The answer is the same safety layer that protects against expensive models getting it wrong: the human in the loop approval framework. Every client facing output gets reviewed by a human regardless of which model produced it. The approval step catches errors from any tier. If a local model's classification sends an email to the wrong bucket, the morning triage report looks wrong and the human catches it. The blast radius of a cheap model error is contained by the same systems that contain an expensive model error.

The Hardware Investment

The entire local AI stack runs on a Mac Studio with an Apple M4 Max chip and 36GB of RAM. Total cost: approximately $2,000 CAD.

On that single machine, we run:

  • Three OpenClaw instances with 20 agents
  • A full RAG database with 33,700+ indexed chunks
  • Multiple local AI models via Ollama
  • 50+ always on background services
  • The complete agent infrastructure (gateway, Slack apps, pollers, schedulers, watchdog)

At an effective agency salary rate of $55/hour, if this machine saves one hour of human work per day (it saves dramatically more), it pays for itself in about 36 business days. Our conservative estimate is that it paid for itself within the first month.

There is no monthly compute bill. There is no cloud hosting fee. There are no scaling costs as we add more agents or more clients. The machine sits on a desk and runs 24/7. When we need to re index the entire knowledge base or process a batch of 10,000 prospect emails, we are not watching a meter tick. We are watching a progress bar on hardware we own.

The Practical Takeaway

If you are building AI agents for your agency and your costs are higher than expected, the answer is probably not a better model. It is a cheaper one, applied to the right tasks.

Run your embeddings locally. Run your classification locally. Run your summarization locally. Run your data cleaning locally. Keep the cloud models for the work that genuinely requires cloud grade intelligence: client facing drafts, complex analysis, and high stakes decisions.

The tools to do this are free. Ollama is open source. Qwen, Llama, and nomic embed text are free to download and run. If you already have a Mac with Apple Silicon (M1 or later) or a Linux machine with a decent GPU, you have the hardware.

The difference between an agent system that costs $5 a day and one that costs $1 a day might not seem like much. But over a year, that is the difference between $1,825 and $365 in AI operating costs. And more importantly, it is the difference between an architecture that scales with your client count and one that scales with your credit card limit.

Your agents do not need to be smart. They need to be cheap where cheap is good enough, and smart only where it matters.

AgencyBoxx runs on dedicated hardware with local AI models handling the heavy lifting and cloud models reserved for high stakes work. Total AI operating cost: about $1 a day. Book a Walkthrough to see the model routing in action.