Six Pillars of Trustworthy Financial AI
Financial AI earns trust only when its reasoning is constrained, inspectable, and replayable. Outside that boundary, it isn’t really a system – it’s uncontrolled behaviour.
Simon Gregory | CTO & Co-Founder
Pillar 1: Auditability
When you can’t see how an answer was formed, you can’t trust it
Pillar 2: Authority
When AI can’t tell who is allowed to speak, relevance replaces legitimacy
Pillar 3: Provenance
When you can’t see the lineage, the system invents it
Pillar 4: Context Integrity
When the evidential world breaks, the model hallucinates the missing structure
Pillar 5: Temporal Integrity
When time collapses, financial reasoning collapses with it
Pillar 6: Determinism
When behaviour is unstable, trust must come from the architecture, not the model
Pillar 3: Provenance
When you can’t see the lineage, the system invents it
Provenance is the continuity layer that preserves the full lineage of information: where it originated, how it was transformed, and whether its meaning survived the journey. It keeps generative systems anchored to their authoritative sources, restoring a principle that has always underpinned trustworthy information: the ability to trace any statement back to its source, its author, and its original context.
It also enables information to be broken into smaller, context preserving units without losing meaning or identity. Without full lineage, any attempt to fragment or recombine information becomes lossy and unreliable.
LLMs break this by default. They don’t reveal what sources they used, how they used them, or whether the output still reflects the author’s intent, even when they appear to. Citations, quotes, and references are a specific sub-classification of the Auditability problem. They are generated artefacts, rather than explicit evidence of actual source usage. Generation is a lossy process, which means the fidelity of original cited content cannot be assumed. The only safe path is to extract the author’s exact words directly from the source, outside the model.
Authority defines who is allowed to speak. Provenance shows which authoritative source was used, how it was transformed, and whether its meaning survived the journey.
The user-side consequence
Without provenance, users cannot validate accuracy, assess authority, inspect context, judge recency, or distinguish high quality sources from low quality ones. The system appears intelligent but shouldn’t be trusted.
The publisher side consequence
Provenance is also the boundary that protects publisher IP. Without it, the LLM interface becomes a value extracting layer that absorbs the publisher’s differentiation and returns none of it. The content still powers the answer, but the publisher becomes invisible.
The model becomes the destination, not the source.
High‑quality content is flattened into generic output. Editorial standards, curation, taxonomy, and expertise are stripped of identity. Engagement shifts from the publisher to the interface. Premium data becomes interchangeable and loses pricing power.
This is the trap created by the rush to deploy flashy LLM interfaces: without provenance, publishers unintentionally disintermediate themselves, and without attribution, they lose visibility. Their IP fuels the system, but the system captures the value.
Provenance reverses this dynamic. It ensures the LLM amplifies the publisher’s value rather than absorbing it. Every answer becomes a pathway back to the authoritative source.
The system level requirement
With provenance, the LLM becomes a precision interface:
- Users can inspect the authoritative source, validate the output against it, and see the lineage of how it was used
- Publishers retain visibility, engagement, and differentiation
- High quality sources are rewarded rather than diluted
- Context survives the generative transformation
- The system becomes transparent, traceable, and trustworthy by design
The human review consequence
Human review cannot compensate for missing provenance. Reviewers are calibrated for human errors (e.g., grammatical slips, broken logic, visible inconsistencies). LLMs fail differently, as they produce truth level errors (e.g., factual misattributions, temporal conflations, hallucinated citations, and subtle semantic drift). These errors are fluent, coherent, and surface-invisible.
Without provenance, a reviewer cannot verify whether a generative answer reflects the underlying source. They cannot check accuracy, confirm intent, or detect invented material. Human review is reduced to a plausibility check rather than a validation step. Provenance is the only mechanism that makes human oversight meaningful.
Relationship to the other pillars
Provenance is the connective tissue between the earlier and later pillars. Auditability requires knowing what the system used and how it used it; provenance exposes both. Authority requires trusted sources; provenance confirms they were invoked. Context Integrity requires preserving meaning; provenance shows whether it survived the generative transformation. Temporal Integrity requires freshness; provenance reveals when the source was published. Determinism requires stable behaviour; provenance provides the fixed inputs and transformations that make stability possible.
Together, these relationships make provenance the structural guarantee that every generative output remains anchored to its authoritative origins, preserving accuracy, context, and economic value. It must be deliberately engineered, because LLMs do not provide it by default and cannot be trusted to reproduce source content safely.
Without provenance, nothing downstream can be trusted.
< Previous | Pillar 2: Authority
Next > | Pillar 4: Context Integrity




