The Engine That Refuses to Lie
Most marketing analytics tools display confident numbers on incomplete data. CDAI does the opposite: when the data isn't trustworthy, the engine flags it and refuses to issue directives. Here's what happened when we ran it on two real businesses with real data quality problems.
The Question Every CFO Asks
"How do I know your tool isn't just hallucinating numbers?"
It's the question that kills analytics deals. Sophisticated buyers — CFOs, in-house data teams, finance committees — have seen too many dashboards confidently display figures that don't survive a five-minute audit. The MarTech category is full of tools that calculate ROAS to two decimal places on data they don't actually have.
CDAI was built specifically against that failure mode. The engine measures one thing — true contribution margin per campaign, after every cost layer the platforms don't report — and issues one of five enforceable directives: SCALE, HOLD, CUT, PAUSE, FLAG.
This document is an operational record. Two real businesses, real data, real engine output. What the engine did. What it refused to do. Why both matter.
What CDAI Measures
Most marketing analytics tools answer the question: "What did this campaign return in revenue divided by ad spend?"
CDAI answers a different question: "What did this campaign actually contribute to the bottom line, after every cost the platforms don't report?"
The Seven-Cost Stack
Platform-reported cost-per-lead and cost-per-acquisition reflect only the costs the platform controls — its own media spend. CDAI reconciles all seven cost layers visible in real lead-gen and acquisition operations:
| Cost Layer | Visible to Google/Meta? |
|---|---|
| 1. Media spend | Yes — the only layer the platforms accurately report |
| 2. Platform fees | Partial — platform-side fees only, not third-party tooling |
| 3. Broker payouts | No — agency margin and lead vendor fees are invisible to platform reporting |
| 4. Refunds | No — post-conversion refund data sits in finance systems, not ad systems |
| 5. Chargebacks | No — payment processor data, weeks-to-months delayed |
| 6. Compliance costs | No — TCPA, HIPAA, state-specific compliance overhead never enters ad reporting |
| 7. Variable costs | No — intake fallout, sales close rate, fulfillment cost |
In real operations, the gap between what platforms report (layer 1, sometimes part of layer 2) and the true cost-per-bound-acquisition can be 30 to 70 percent. The campaign that looks profitable on a Google Ads dashboard is often the campaign that's quietly destroying margin once the full cost stack is reconciled.
CDAI exists to close that gap. The rest of this document is what happened when it ran on real data.
Test Case 1: E-Commerce Business
The Test
The business ran a Google Search Ads campaign from August 31, 2025 to December 2, 2025. The campaign was complete (concluded), with full historical data exportable from Google Ads.
What the Engine Did
This confirms what every analytics tool claims but few actually demonstrate: that the engine can read real exported platform data, not just demo fixtures.
directive_safe.
On this data, the monitor identified the campaign as approximately four months stale and set
directive_safe = FALSE.
The engine refused to issue directives.
This is the behavior most analytics tools do not have. A SCALE directive issued on stale data is worse than no directive at all — it tells a marketer to commit budget to a campaign whose underlying conditions may have changed. CDAI's health monitor exists specifically to prevent that failure mode.
This is the architectural foundation that allows CDAI to safely serve multiple clients from the same engine. It was verified on real data, not just simulation.
• CDAI ingests real exported platform data without modification or fabrication
• The engine calculates against whatever cost layers are populated and does not invent values for layers that aren't
• The health monitor functions as a safety gate — directives do not issue on data the engine cannot stand behind
• Multi-tenant data isolation works on real data, not just architecture diagrams
Test Case 2: Healthcare Services Business
The Test
The business provided 3.5 months of Meta Ads Manager exports from January 1, 2026 to April 22, 2026, plus a CRM contacts file. The data set represents a real-world condition: messy, partial, and with one critical attribution gap.
What the Engine Did
Cross-tenant isolation verified: The data was inserted into the same physical database that already contained the first organization's data. Subsequent queries filtering by org_id returned only this organization's data. The architectural isolation works on real client data — not just in test fixtures.
This is a discipline most ingestion pipelines abandon at scale. When a CSV has a missing column, the easy path is to fill it with a sensible default and keep going. CDAI does not. If the data isn't there, it isn't there.
"Contact Import" — meaning no campaign attribution existed at the lead level. There was no way to determine which Meta campaign generated which lead.
Without that attribution, the engine cannot calculate true cost-per-lead by campaign, contribution margin by campaign, or issue SCALE / HOLD / CUT / PAUSE / FLAG directives at the campaign level. These are the engine's primary outputs.
The engine did not attempt to fabricate the attribution it lacked. Lead records were not ingested. The audit pipeline halted at the data quality gate.
A weaker tool would have done one of three things: (a) ingested the leads with null campaign_id values and silently degraded the analysis, (b) used a heuristic — chronological proximity, last-touch fuzzy match — to invent attribution, or (c) issued directives anyway, qualified by a footnote. CDAI did none of those. The audit was incomplete because the data was incomplete, and the engine reported that condition rather than disguising it.
directive_safe = FALSE. No directives were issued.
This is the same behavior demonstrated on the first test. The trigger condition was different (stale data on Test 1; missing attribution on Test 2), but the engine's response was the same: refuse to issue directives the data cannot support.
• Multi-tenant isolation works on real client data, in production, against a database that already contains other organizations
• The engine does not fabricate attribution to fill gaps — it surfaces the gap as a finding
• The health monitor responds to multiple distinct integrity failure modes, not just one
• CDAI's data quality discipline is structural, not optional
What These Two Tests Prove Together
Two real businesses. Two different industries. Two different platforms (Google Ads and Meta Ads). Two different data quality problems (stale data and missing attribution). One consistent engine behavior.
| Test Dimension | Test 1 | Test 2 |
|---|---|---|
| Platform | Google Ads | Meta Ads |
| Industry | E-commerce | Healthcare / senior care |
| Data Condition | Complete but ~4 months stale | Recent but missing campaign attribution |
| Ingestion | Successful | Successful (campaigns + costs); halted on lead attribution gap |
| Multi-Tenant Isolation | Verified | Verified |
| Health Monitor Result | directive_safe = FALSE (stale) | directive_safe = FALSE (incomplete) |
| Directives Issued | None — by design | None — by design |
| Fabricated Values | Zero | Zero |
The Behavior Worth Naming
CDAI is built on a principle that's rare in the marketing analytics category: the engine refuses to issue an output it cannot stand behind.
That sounds obvious. In practice, it's not. The competitive landscape is full of tools that confidently display modeled conversions, attributed revenue, and projected ROAS — calculated on data with significant integrity issues, presented without any indication that the underlying inputs were incomplete or stale. Most CFOs evaluating these tools have learned to discount the headline numbers by 30 to 50 percent before believing them.
The two tests above demonstrate the opposite posture. When the data was complete enough but stale, the engine flagged it and refused to issue directives. When the data was recent but missing critical attribution, the engine surfaced the gap and refused to invent the attribution. In neither case did CDAI produce a confident-looking output that wouldn't have survived scrutiny.
That refusal-to-fabricate behavior is the single most important property an analytics engine can have when the buyer is a CFO with budget authority and a long memory for tools that lied to them.
What CDAI Looks Like With Complete Data
So that the integrity-first picture above isn't mistaken for a limitation, here's what CDAI's full output looks like when the cost stack is populated and the data is current.
In a separate prior test, the engine was run against a complete simulated data set — six campaigns, full seven-layer cost stack populated, all data current. The engine executed end-to-end without error and produced one directive per campaign across all five directive types:
- SCALE — for campaigns where true contribution margin justified increased spend
- HOLD — for campaigns with developing data not yet conclusive
- CUT — for campaigns where true margin was negative once the full cost stack was reconciled
- PAUSE — for campaigns showing anomalies under review
- FLAG — for campaigns requiring manual review
Each directive carried a numeric confidence score and a reason code. The directive sheet is the engine's primary client-facing artifact and is what every paying audit will deliver.
The simulated test confirms the engine's full output capability. The two real-data tests above confirm the engine's behavior under real-world data quality conditions. Both are necessary. Neither is sufficient on its own.
Why This Matters to a Buyer
If you run paid acquisition at any meaningful scale — six figures monthly or more — three things follow directly from these tests.
One · Your reported cost-per-acquisition is structurally incomplete
The platforms report what they control. They do not reconcile broker margin, refunds, chargebacks, or compliance overhead. A campaign that looks profitable in your Google Ads or Meta dashboard is often the campaign that's destroying margin once the full cost stack is reconciled. The 30 to 70 percent gap between reported CPL and true cost-per-bound-acquisition is not a hypothesis — it's structural arithmetic that compounds every month you don't measure it.
Two · The tools that say they fix this often don't
The contribution-margin-aware tools in market today are largely Shopify-native, optimized for ecommerce SKU economics. They do not handle lead-gen broker payouts, regulated-industry compliance overhead, multi-platform partner ecosystems, or refund-and-chargeback dynamics specific to high-ticket services. CDAI is built for these conditions specifically.
Three · The engine you choose has to refuse to lie
If your CFO is going to make capital allocation decisions on the engine's output, the engine has to be willing to tell you when it doesn't have enough data to recommend an action. CDAI does this. The two tests above are the proof.
The Question That Reveals the Gap
The people to think about are anyone who runs or owns:
- A personal injury or mass tort law firm — spending on lead vendors like EvenUp, Axiom, Litigation Leaders, or running their own Google/Meta campaigns
- An insurance agency or Medicare brokerage — buying leads from aggregators or running direct acquisition
- A senior care or assisted living facility — acquiring patients through lead vendors, referral networks, or paid ads
- A home services company — roofing, solar, HVAC, water damage restoration — buying HomeAdvisor, Angi, Thumbtack leads or running their own campaigns
- Any business spending serious money buying leads from vendors — if they're writing five-figure monthly checks to lead sources and can't tell you the true cost per customer that actually closes and stays closed
See What Your Platform Isn't Telling You
Most businesses are making capital allocation decisions on numbers that are wrong by 30–70%. The gap between what your dashboard reports and what you're actually paying is structural arithmetic — it compounds every month you don't measure it.
A 30-day distortion audit on your campaign data costs $3,500 and delivers within 7 days. If we don't surface margin distortion you weren't tracking, you don't pay.
Request a Distortion AuditTechnical Architecture Summary
The CDAI Engine is deployed on:
- Database: Supabase (PostgreSQL) with Row-Level Security policies
- Backend: Python on Render
- Portal: React on Vercel (live at cdai-portal.vercel.app)
- Scheduler: Deployed and running automated health checks
- Email Delivery: Operational via Resend integration
Multi-tenant capable. Zero consumer PII stored in any table.