AI-generated insights have a trust problem. When a dashboard says "MRR is $287,000," people believe it — because a human wrote the SQL, a human reviewed the logic, and a human signed off on the dashboard.
When an AI says "MRR is $287,000 and churn is likely to spike next month because of 3 enterprise accounts with declining usage," the immediate reaction is: how do you know? What data did you use? How confident are you? Can I trace this back to the source?
If you can't answer those questions, the insight is useless. Not because it's wrong, but because nobody will act on it.
The confidence score framework
Every insight Vesh AI generates carries a confidence score between 0 and 1. This score isn't a single number — it's a composite of four independent confidence dimensions:
Source confidence (how reliable is the data?)
Not all data sources are equal. Stripe billing data is highly reliable — it's the system of record for payments. Product usage data from your database is moderately reliable — it depends on your instrumentation quality. CRM data is lower reliability — it's manually entered and often stale.
We assign source confidence weights based on the type of data and the freshness of the last sync: - Billing system (Stripe): 0.95 - Product database (direct connection): 0.85 - CRM (API sync): 0.70 - Manual import: 0.50
These weights are configurable. If your CRM data is exceptionally well-maintained, you can adjust the confidence upward.
Entity confidence (how sure are we about identity?)
Entity resolution produces match scores for every pair of linked records. If a Stripe customer and a Postgres user share an identical email address, the entity confidence is 0.98. If they match only on a fuzzy company name, it might be 0.72.
The entity confidence for an insight is the minimum match confidence across all entities involved. If an insight references 5 entities and one has a weak match (0.65), the entity confidence is 0.65 — because the weakest link determines overall reliability.
Metric confidence (how trustworthy is the calculation?)
Metric confidence depends on data completeness and computational accuracy: - Were all expected source records available? (Coverage) - Did the computation encounter any edge cases? (Anomalies in input) - Is the metric value within expected bounds? (Sanity check) - How many source records contributed? (Sample size)
A metric computed from 500 entities with complete data gets a higher confidence than one computed from 15 entities with missing fields.
Anomaly confidence (how significant is the finding?)
For anomaly-based insights, we compute statistical significance using z-scores against the historical baseline. An anomaly with a z-score of 8.5 (extremely unlikely to be random) gets a higher confidence than one with a z-score of 2.1 (might be noise).
We also factor in whether the anomaly has a plausible causal explanation. An MRR spike that decomposes cleanly into "3 enterprise upgrades" is more trustworthy than one with no clear attribution.
Full data lineage
Confidence scores tell you how much to trust an insight. Lineage tells you why — by making the entire computation traceable from insight back to source records.
Every insight in Vesh AI includes a lineage trace:
You can click through this chain in the admin UI. If something looks wrong, you can trace it all the way back to the raw Stripe record that caused it.
Human-in-the-loop feedback
Confidence scores and lineage make insights inspectable. But the final trust-building mechanism is feedback.
Every insight delivered to Slack includes two buttons: thumbs up and thumbs down. When a user marks an insight as incorrect, three things happen:
- . The insight is flagged for review, and its confidence is retrospectively adjusted
- . The feedback is stored and used to calibrate future confidence scoring for similar patterns
- . If a specific data source consistently produces low-confidence or rejected insights, it's flagged for the admin to investigate
Over time, the system learns which types of insights are trusted and which aren't — and adjusts its thresholds accordingly.
Why this matters for adoption
The difference between a tool that gets used and one that gets ignored is trust. Dashboards get used because people built them and understand them. AI insights get ignored because they feel like black boxes.
Confidence scores, lineage, and feedback turn the black box into a glass box. The AI still does the computation — entity resolution, metric calculation, anomaly detection, causal decomposition — but every step is visible, traceable, and correctable.
That's what makes the difference between a toy and a tool. Not the accuracy of the AI (which is necessary but not sufficient), but the ability of the human to verify, trust, and act on what the AI produces.
We believe this trust layer is the missing piece in AI analytics. Not better models, not more data, not fancier visualizations — just transparency about how conclusions were reached and confidence about how reliable they are.