I spent 6 years doing this work manually for 100+ brands. Then I built the tool that would have replaced me. 61 seconds vs 5 business days, real output comparison, no sugarcoating.
I Know What Agencies Charge For. I Used to Be the One Charging It.
For 6 years, I sat in the agency chair. Performance marketing. $10M+ in managed ad spend across Meta, Google, and TikTok. Over 100 DTC brands. I know exactly what the $5,000/month retainer buys, because I used to be the one delivering it.
Here's what the monthly deliverable typically looked like:
- Pull screenshots from Ads Manager (20 minutes)
- Drop them into a slide template (30 minutes)
- Write a paragraph of "insights" that say nothing specific (45 minutes)
- Add a recommendations slide that says "optimize underperforming campaigns" (15 minutes)
- Send it on Friday so the client thinks you worked on it all week
Total real work: about 2 hours. Billed at: $5,000.
I'm not proud of this. I'm being honest. And if you've ever worked at a performance marketing agency, you know this is accurate. The uncomfortable truth is that the actual analysis, the part where you stare at data and form opinions about what to change, takes a fraction of the time clients think it does. The rest is packaging.
Last month, I decided to run an experiment. I took a real client account, a DTC ecommerce brand spending roughly $5,000/week across Meta and Google, and gave the exact same brief to two contestants:
Contestant A: A human agency team (the kind I used to lead).
Contestant B: Three AI agents I built, running simultaneously on the same data.
Same brief. Same data. Same deadline. Let's see who wins.
The Brief
Simple. The kind of thing agencies get every month:
"Audit the account. Tell me where we're wasting money, flag creative fatigue, and give me a 2-week action plan with specific budget recommendations."
This is bread-and-butter agency work. If you can't nail this, you can't justify the retainer.
Contestant A: The Agency (5 Business Days)
The deck arrived on Friday. 12 slides.
Slide 1: Cover page with the client logo.
Slide 2: Agenda ("Account overview, performance summary, recommendations").
Slide 3-5: Screenshots from Ads Manager with arrows pointing at numbers I can read myself.
Slide 6: "Performance Summary": "Overall performance was mixed this period. Some campaigns performed well while others require optimization."
Slide 7-8: Two graphs that restated the screenshots as bar charts.
Slide 9: "Recommendations": "We suggest increasing budget on top-performing campaigns and testing new creative for underperforming placements. Consider diversifying channel strategy."
Slide 10: "Next Steps": "Review creative assets. Schedule optimization call."
Slides 11-12: Thank you page and contact details.
Not a single campaign was named in the recommendations. Not a single dollar amount suggested for reallocation. Not a single ad flagged for creative fatigue with an urgency level. Not a single timeline attached to any action.
I know this deck. I've written this deck. I've written this deck a hundred times.
Time to deliver: 5 business days.
Contestant B: Three AI Agents (61 Seconds)
I run Cresva, an AI marketing intelligence platform I built solo from India. It has 7 specialized agents:
- Parker: Attribution and budget integrity
- Olivia: Creative intelligence and fatigue detection
- Sam: Strategic planning and growth opportunities
- Felix: Forecasting and trend prediction
- Maya: Brand memory and institutional knowledge
- Dana: Data quality and health monitoring
- Dex: Reporting and delivery
The orchestrator read the brief and routed it to Parker + Olivia + Sam, the three agents most qualified for an account audit.
Phase 1: Prefetch (2.9 seconds) The orchestrator pulled every metric in a single parallel batch: campaigns, ad sets, ads, creative data, device breakdowns, placement data, daily spend, Shopify revenue. One fetch. Everything cached.
Phase 2: Agent Execution (25 seconds) Three agents ran simultaneously. Parker analyzed budget allocation and attribution gaps. Olivia scanned every creative for fatigue signals. Sam identified strategic opportunities and risks. They didn't hit the database once, they reasoned over the pre-loaded data through lightweight tool handlers.
Phase 3: Synthesis (30 seconds) Claude Sonnet took all three analyses and wrote a unified strategic report. Not a merge, a synthesis. Where Parker and Sam agreed on a budget recommendation, confidence went to HIGH. Where Olivia's creative data conflicted with Sam's spend suggestion, the report flagged it as a tension and explained the tradeoff.
Total: 61 seconds.
Here's what came out.
The Executive Summary
What the AI wrote:
"The account is burning $1,700 weekly on a Meta prospecting campaign delivering just 1.6x ROAS while Google brand search sits underfunded at 11x+ ROAS with only 8% of total budget. Immediate reallocation can recover an estimated $3,500 in weekly revenue while maintaining current spend levels."
What the agency wrote:
"Performance was mixed this period with some campaigns performing well and others requiring optimization."
Read those side by side. One sentence identifies the exact problem, quantifies the waste, and estimates the recovery. The other says nothing. Literally nothing. You could paste that sentence into any account in any industry and it would be equally meaningless.
I've written the agency version. I've written it at 4pm on a Friday when I needed to send something before the weekend. The AI doesn't have Fridays. It doesn't have shortcuts. It just analyzes.
Budget Recommendations
The agency said: "We suggest increasing budget on top-performing campaigns."
The AI said:
Google Brand Search campaign: Current: $180/week. Recommended: $400/week. Change: +$220. Expected Impact: +$8,000 monthly revenue. Confidence: HIGH.
Meta Prospecting campaign: Current: $1,700/week. Recommended: $1,300/week. Change: -$400. Expected Impact: Maintain revenue at lower cost. Confidence: MODERATE.
Meta BAU campaign: Current: $1,050/week. Recommended: $700/week. Change: -$350. Expected Impact: Reallocate to higher ROAS channels. Confidence: MODERATE.
But here's what made me stop scrolling, the anti-recommendations:
"Do NOT scale brand search purely for ROAS. It captures existing demand, not new customers. 11x+ ROAS on low weekly spend likely reflects a small sample where 1-2 lucky conversions inflate the number. Validate with 2 weeks of increased spend before committing further."
That's the kind of nuance I'd expect from a senior strategist with 5+ years of experience. The AI questioned its own recommendation. It flagged the risk in its own analysis. An agency that recommends "increase budget on top performers" without caveats is giving you a way to waste money confidently.
Creative Fatigue Analysis
The agency wrote one bullet: "Some ads may be experiencing fatigue. We recommend testing new creative."
Olivia wrote:
Ad 1: Static image, product benefit angle. 12 days active. CTR 0.27%, declining. Frequency 4.2. Verdict: Replace this week.
Ad 2: Carousel, price comparison. 9 days active. CTR 0.27%, stable. Frequency 3.5. Verdict: Can run 2 more weeks.
Ad 3: Video, subscription offer. 14 days active. CTR 0.31%, declining. Frequency 3.8. Verdict: Replace next week.
And then the prediction: "Ads with CTR below 0.3% at current frequency will hit fatigue threshold within 5 days."
As someone who spent years manually checking frequency caps and CTR trends in Ads Manager, usually on the day before a client call, watching an AI do it in seconds with predictions attached felt like watching someone automate my entire Tuesday afternoon.
The 2-Week Action Plan
The agency gave me: "Review creative assets. Schedule optimization call."
The AI gave me:
Day 1-3: Immediate
- Double Google brand search budget from $180 to $400/week
- Reduce Meta prospecting campaign by $400/week
- Pause the static image ad (4.2 frequency, declining CTR)
- Launch 2 new creatives to replace fatigued ads
Day 4-7: This Week
- Monitor brand search impression share, target 85%+ without bid inflation
- Reduce Meta BAU campaign by $350/week if Google scaling succeeds
- Test desktop-optimized creatives (desktop showing 21% higher ROAS than mobile)
- Set automated rules to pause ads at 4.0+ frequency
Week 2: Next Week
- Review brand search volume ceiling, cap budget at effective level
- Assess revenue impact from Meta reductions
- Replace subscription video ad before hitting 4.0 frequency
- Scale winning creatives from Week 1 testing
Every action names a specific campaign type. Every action has a specific number. Every action has a day.
When I worked at agencies, action plans looked like the agency's version because nobody had time to be this specific. You're managing 8-12 accounts. You have 2 hours per account per week. Being specific is a luxury. The AI doesn't have a client load. It has one account and unlimited attention.
Then I Pushed It Further: Deep Research
The audit was Contestant B's first round. I wanted to see if it could do what agencies take 2-3 weeks for.
The brief: "Analyze the competitive landscape in our vertical. Who are the top 5 DTC competitors, what channels are they scaling on, and where are the gaps we can exploit?"
3 minutes and 47 seconds later, a research report arrived. 146 web sources cross-referenced against the client's internal data.
The competitive table mapped 5 real competitors with their channel splits, pricing strategies, estimated ROAS, key strengths, and vulnerabilities. Not generic advice. Real companies. Real data points.
The performance benchmark table showed the client vs industry averages vs category-specific benchmarks, with a status rating on each metric. Google ROAS: "Exceptional, top 5% of category." Meta ROAS: "Below category leaders."
The killer insight: top-performing brands in the category allocate 40-50% of budget to Google Search, 25-30% to short-form video, and only 25-35% to Meta retargeting. The client's split was 81% Meta, 19% Google, almost the inverse of what market leaders were doing.
The report concluded that the client was leaving thousands in weekly revenue on the table by underfunding their most efficient channel while overspending on their least efficient by double digits.
That's not a report. That's a strategic reframe that would take a senior strategist a week to research and articulate. It took 3 minutes and 47 seconds.
The Scorecard
Time to deliver audit. Agency: 5 business days. AI: 61 seconds.
Time for competitive research. Agency: 2-3 weeks. AI: 3 minutes 47 seconds.
Monthly cost. Agency: $5,000. AI: $999 platform plus roughly $50 in API costs.
Campaigns named in recommendations. Agency: 0. AI: Every single one.
Dollar amounts in budget recommendations. Agency: 0. AI: Exact figures per campaign.
Anti-recommendations included. Agency: No. AI: Yes, with risk explanations.
Sources cited in research. Agency: 0. AI: 146.
Second-order thinking. Agency: No. AI: Sample size flags, brand vs prospecting ROAS distinctions, frequency decay predictions.
Winner: not close.
The Technical Bit (For Builders)
Three architectural decisions that make the 61-second time possible:
Multi-model orchestration: GPT-4o handles data retrieval and tool calling. Fast, cheap, reliable at structured tasks. Claude Sonnet writes the final synthesis. Better at strategic reasoning, nuance, and structured recommendations. Each model does what it's best at.
Prefetch-and-inject pattern: Instead of 3 agents each making 10-16 database queries (30-55 total), the orchestrator fetches all data once in a single parallel batch. Agents reason over pre-loaded data through lightweight shim handlers. Database goes from 55 queries to zero during agent execution.
Circuit breakers: If Claude's API fails, GPT-4o writes the synthesis. If the database drops connections, Redis serves cached data with freshness flags. If web research fails, the report continues with internal data and flags the gap. The user never sees a broken session.
Full architecture deep dive in my previous post: "I Built 7 AI Agents That Run Marketing Operations. Here's the Entire Architecture."
What This Actually Means
I need to be honest here. Because I'm not some outsider taking shots at agencies. I was the agency. I sat in that chair for 6 years.
What AI automates today:
- Account audits and performance analysis
- Budget reallocation recommendations
- Creative fatigue detection and prediction
- Competitive benchmarking and market research
- Channel mix optimization
- Day-by-day campaign action plans
What still needs humans:
- Relationships with platform reps who give you early access to betas and credits
- Creative production: the actual video shoots, design work, copywriting
- Brand strategy that requires cultural context and taste
- Client management and communication
- The gut call when data says one thing but experience says another
- Negotiating rates with publishers and media partners
The uncomfortable truth: if you strip away the relationship layer and the creative production, the core analytical work that agencies bill for, looking at dashboards, pulling insights, writing recommendations, that work is now automatable at higher quality, higher speed, and 1/50th the cost.
I built the tool that would have replaced me 3 years ago. That's a strange thing to sit with. But I'd rather be the one building the replacement than the one pretending it's not coming.
The Question Nobody Wants to Ask
There are roughly 120,000 digital marketing agencies worldwide. The majority of their recurring revenue comes from reporting, analysis, and strategic recommendations. Exactly the work that three AI agents just did in 61 seconds.
The question isn't whether AI will replace agencies. That's already happening. The question is: what happens to the $400 billion digital advertising industry when the analysis layer becomes essentially free?
What does an agency look like when its value shifts entirely to creative production, relationship management, and strategic taste? What does a performance marketer's career look like when the work they trained for can be done by software?
I don't have answers. But I have 3 AI agents, 61 seconds, and the uncomfortable knowledge that I built the thing that makes my old job unnecessary.
That's either terrifying or exciting. Depends which chair you're sitting in.
