TrueCase

AI-powered business case builder for AI investments — generates governance-adjusted ROI projections with a reliability score, UK regulatory citations, and a Claude-powered three-section narrative grounded entirely in a verified knowledge base.

Shipped• By Ogbebor OsaheniApril 2026

Executive Summary

TrueCase bridges the gap between financial optimism and regulatory reality in AI investment decisions. A four-stage guided flow — contextual questionnaire, governance assessment gate, financial projections, and AI narrative export — produces a business case built on two projections: the headline ROI and the governance-adjusted ROI. The gap between them is the product. Zero user data is retained. Every governance claim traces to a primary UK regulatory source. Every Claude-generated sentence is constrained to a verified knowledge base — not Claude's training data.

6

Governance elements

Each KB-backed with UK regulatory citations and weighted reliability deductions

3

Narrative sections

Business Case Summary, Governance Assessment, Social Return Signal

0

Bytes retained

Stateless by design — no database, localStorage, cookies, or logging

3

Knowledge bases

governance-kb.json, benchmarks.json, sdg-kb.json — all verified April 2026

Problem Statement

Business cases for AI investments tend to be optimistic. Governance gaps — missing human oversight, undocumented training data, no audit trail — are treated as implementation details rather than financial risks. The result is a board deck that shows £2.4M annual gain while quietly ignoring that the system has no regulatory compliance map and may trigger a GDPR enforcement action that erases that gain entirely.

Finance teams want ROI projections. Compliance teams want governance verification. These conversations happen in separate rooms, in different languages, with different tooling. There is no shared instrument that lets both teams look at the same number and ask: 'How confident should we be in this?'

The challenge is not technical — ROI calculators exist, and compliance checklists exist. The barrier is product thinking: how do you build a tool that makes governance maturity visible as a financial variable, not a post-hoc audit finding? How do you ground AI-generated narrative in verified regulatory sources rather than confident-sounding hallucination? How do you do this without storing any of the sensitive business data the user just typed in?

"The governance gap between headline ROI and governance-adjusted ROI is not a failure state — it is the most valuable thing TrueCase shows. When a user's reliability score is 45% because three governance elements are missing, their real ROI projection is 45% of what they thought it was. That is not a discouragement. That is a roadmap."

User Personas

TrueCase is designed for three distinct user archetypes, each arriving at the tool with a different job-to-be-done. The product must serve all three without asking them to self-identify — the guided questionnaire (Q1: decision type, Q4: sector) routes each user to the right regulatory context automatically.

I

Internal AI PM

AI Product Manager seeking budget approval at a UK enterprise

Age 34 • Tech: High

Goals:

  • Build a board-ready business case that finance and compliance will both sign off
  • Anticipate governance objections before the presentation, not during it
  • Show that the AI investment has been assessed against UK regulatory requirements

Pain Points:

  • Finance wants a number. Compliance wants a framework. They are never in the same room.
  • ROI calculators give optimistic projections but no governance context — useless for regulated sectors
  • Writing business cases from scratch is slow; writing defensible ones is slower

"My CFO will ask what happens if this triggers a GDPR audit. My answer right now is 'we'll deal with it.' I need a better answer before next Thursday's board meeting."

S

Startup Founder

Founder pitching institutional investors or enterprise buyers for an AI product

Age 30 • Tech: High

Goals:

  • Show investors that governance risk has been identified and mitigated, not ignored
  • Produce a formal, credible ROI document that survives a due diligence review
  • Demonstrate regulatory awareness to enterprise buyers in financial services or healthcare

Pain Points:

  • Investor due diligence now routinely asks about AI governance — founders are unprepared
  • Enterprise buyers in regulated sectors will not sign without evidence of compliance mapping
  • No time or budget to commission a formal governance audit at seed or Series A stage

"The VC asked me about our governance maturity. I didn't have a good answer. I need to be able to show that I've thought about this seriously before the next meeting."

E

Enterprise Buyer

Technology Director or Procurement Lead stress-testing an AI vendor's ROI claim

Age 47 • Tech: Medium

Goals:

  • Audit a vendor's ROI projection against independent governance criteria
  • Produce a governance gap assessment as part of the procurement process
  • Justify the purchase (or rejection) to legal, compliance, and the board

Pain Points:

  • Vendor ROI claims are always optimistic and never mention governance dependencies
  • No neutral tool exists to run a vendor's numbers through an independent governance filter
  • Procurement decisions in regulated sectors carry personal liability — need defensible documentation

"Every vendor says their AI will save us £1.2M a year. None of them tell me that's only true if we have a Human Review Requirement and a Decision Audit Trail in place — which we don't."

Solution Overview

TrueCase is a four-layer product that flows top to bottom: a contextual questionnaire sets regulatory exposure; a governance assessment gate scores maturity against six KB-backed elements; financial projection cards show headline and reliability-adjusted ROI side by side; and a Claude-generated three-section business case narrative — grounded in verified knowledge base entries, not Claude's training data — is exported as PDF or HTML. Nothing is stored. Every claim is traceable.

Contextual Questionnaire (4 Questions)

Decision type, data type, scale, and sector questions route users to the right governance context — triggering sector-specific regulatory anchors, benchmark ranges from benchmarks.json, and auto-confirmation of Element 6 when no personal data is involved.

Governance Assessment Gate

Six governance elements from governance-kb.json, each with a 3-state toggle (Confirmed / Partial / Missing), regulatory citations sorted by sector priority, consequence text from the KB, and weighted deductions that compose the reliability score (0–100%).

Reliability-Adjusted Projections

Six financial output cards, always rendered with headline and reliability-adjusted gain at identical size. The governance gap — the difference between the two — is always visible. Not a warning. A number.

AI Narrative + Export

Claude Sonnet 4 generates three sections (Business Case Summary, Governance Assessment, Social Return Signal) grounded in governance-kb.json, benchmarks.json, and sdg-kb.json. Exported as a 3-page PDF or standalone HTML file. All inputs discarded immediately after generation.

Architecture Diagram

Data & Methodology

Data Dictionary

FeatureTypeDescriptionSource
hard_output_capboolean (confirmed / partial / missing)Hard, non-configurable maximum on AI outputs. Missing: -25 reliability points. Partial: -12. The single highest-weighted element — output without a ceiling can cause catastrophic harm at scale.governance-kb.json (UK AI Act signals, ICO guidance)
human_overrideboolean (confirmed / partial / missing)Human review required before any AI decision takes effect. Missing: -20 reliability points. Partial: -10. Automation without override creates accountability gaps under GDPR Article 22.governance-kb.json (GDPR Article 22, FCA Consumer Duty)
explainabilityboolean (confirmed / partial / missing)Decision audit trail with plain English explanation available within 24 hours of request. Missing: -15 reliability points. Partial: -7. Required for right-to-explanation obligations.governance-kb.json (GDPR Article 22, Ofcom Broadcasting Code)
regulation_identifiedboolean (confirmed / partial / missing)All applicable UK regulations identified and formally mapped. Missing: -20 reliability points. Partial: -10. Unmapped regulation creates silent liability that compounds over time.governance-kb.json (sector-specific: FCA, Online Safety Act, ICO)
training_data_documentedboolean (confirmed / partial / missing)Training data source, known limitations, and drift detection plan documented. Missing: -10 reliability points. Partial: -5. Undocumented data provenance is a due diligence failure in any regulated sector.governance-kb.json (ICO AI Auditing Framework)
personalisation_boundaryboolean (auto-confirmed if Q2 ≠ personal)Policy defining where personalisation ends and surveillance begins. Missing: -10 reliability points. Partial: -5. Auto-confirmed if user indicated no personal data in Q2 — only applies to personal data processing.governance-kb.json (UK GDPR, ICO guidance on profiling)

Methodology

The reliability score is computed as a weighted deduction sum from governance-kb.json. It starts at 100% and subtracts the element's reliability_reduction_missing for each missing element, or reliability_reduction_partial for each partial. The final score ranges from 0% (all elements missing) to 100% (all confirmed). This score is then applied as a multiplier to the headline ROI projection: Reliability-Adjusted Gain = Projected Gain × (reliability_score / 100). The reliability score is not a compliance grade — it is a financial variable. A 60% score on a £2M projected gain means the governance-adjusted figure is £1.2M. The £800k gap is the governance risk made visible as money. All financial benchmark ranges in benchmarks.json are sourced from primary UK regulatory documents (legislation.gov.uk, FCA reports, Ofcom research) and marked as 'verified' or 'illustrative range' with source URLs. SDG mappings in sdg-kb.json are derived from UN SDG target documentation and cross-referenced against Q1 (decision type) and Q4 (sector) combinations.

Validation Approach

  • All regulatory anchors in governance-kb.json link to legislation.gov.uk or official regulatory body publications — no secondary sources
  • Sector-specific benchmark ranges in benchmarks.json sourced from primary FCA reports, Ofcom research, and GDPR enforcement records — all marked with source URL and verification status
  • Claude system prompt explicitly constrains output to KB entries and user inputs: banned from adding claims beyond verified sources, banned from 'leveraging', 'synergies', or other filler assertions
  • Self-assessment disclaimer always visible: 'TrueCase cannot verify these answers' — no false authority claimed for governance gate outputs
  • PDF and HTML exports include KB version, generation date, and liability disclaimer on every page

Ethics & Responsible AI

TrueCase's ethical commitments are architectural, not aspirational. Privacy by design means the system physically cannot retain data — there are no write operations to a persistence layer. Grounded generation means Claude cannot add governance claims beyond what the verified KB and user inputs supply. Self-assessment transparency means the reliability score is always accompanied by a disclaimer that TrueCase cannot verify the answers the user provided.

Privacy by Design

No database, localStorage, sessionStorage, cookies, or server-side logging. The PDF API route processes inputs, generates the file stream, and discards all data immediately. Stateless architecture is enforced at code level — not policy level.

Grounded Generation

Claude Sonnet 4's system prompt explicitly prohibits adding factual claims beyond governance-kb.json and user inputs. Regulatory citations must come from KB entries. Financial figures must come from user inputs. SDG mappings must come from sdg-kb.json. There is no fallback to training data.

Self-Assessment Transparency

The grounding disclosure is always visible above the generate button. The AI-generated label is always shown after narrative output. The self-assessment disclaimer ('TrueCase cannot verify these answers') appears at the reliability score. Three separate honesty markers, none optional.

Uniform Scoring

The reliability score formula is deterministic and public: start at 100%, subtract KB-specified deductions per element status. No demographic data used. No sector weighting in the score calculation. Identical governance configurations produce identical reliability scores regardless of who is using the tool.

Guardrails & Safeguards

RuleThresholdRationale
Claude Output ScopeKB + user inputs onlyPrevents hallucinated governance claims. Every factual assertion in the narrative must trace to governance-kb.json, benchmarks.json, sdg-kb.json, or the user's own inputs.
Data Retention0 bytes per sessionNo write operations to any persistence layer. PDF and HTML generation uses inputs only within the server-side request lifecycle — all discarded on response.
Regulatory Citation SourcePrimary sources onlyAll UK regulations cited to legislation.gov.uk or official regulatory body publications. Secondary sources and general assertions excluded from KB at verification stage.
Narrative Banned LanguageProhibited: 'leveraging', 'synergies', 'driving value', 'unlock potential'Filler corporate language without substantive meaning is explicitly banned in the Claude system prompt — reinforcing the product's credibility commitment.

Bias Audit & Fairness Assessment

No user demographic data is used in any calculation. The reliability score is a function of governance element statuses only — identical inputs produce identical scores for all users. Sector-specific benchmark ranges in benchmarks.json are sourced from public regulatory data (FCA, Ofcom, ICO) and apply uniformly within each sector. The SDG mapping logic in sdg-kb.json assigns impact categories based on Q1 (decision type) and Q4 (sector) — structural use-case characteristics, not user identity. The self-assessment disclaimer ensures no authority is falsely attributed to user-provided governance confirmations.

Risk Register

TrueCase's risk profile is shaped by two primary tensions: the knowledge base is static but UK AI regulation is evolving rapidly, and the governance assessment relies entirely on user self-reporting which creates honest score inflation as a vector. Both risks were identified at design stage and addressed architecturally.

RISK-001HIGHKNOWLEDGE / REGULATORY

UK AI Regulation Evolves Faster Than KB Versioning Cycle

16
L:4 × I:4
Mitigation: KB entries carry a last_verified timestamp (April 2026). PDF and HTML exports include the KB version and generation date on every page. Users are aware the KB reflects a specific regulatory snapshot. Quarterly review cadence planned for KB maintenance.
Residual Risk: 8 (reduced from 16)Owner: Product Manager
RISK-002HIGHTRUST / ASSESSMENT INTEGRITY

Users Self-Report Governance Controls That Are Not Actually In Place

12
L:4 × I:3
Mitigation: Self-assessment disclaimer ('TrueCase cannot verify these answers') always visible at the reliability score. AI-generated disclaimer in all exports makes the self-assessment basis explicit. Product is designed for honest users — it is a planning tool, not an audit instrument.
Residual Risk: 8 (reduced from 12)Owner: Product Design
RISK-003HIGHTECHNICAL / AI OUTPUT

Claude Outputs Governance Claims Beyond Verified KB Sources

10
L:2 × I:5
Mitigation: System prompt explicitly constrains Claude to KB entries and user inputs. Banned language list prevents filler assertions. Grounding disclosure and AI-generated label accompany every output. Users are told the source of every claim.
Residual Risk: 4 (reduced from 10)Owner: Product Manager
RISK-004HIGHPRIVACY / TRUST

User Distrust of Data Handling in a Tool Handling Sensitive Business Cases

12
L:3 × I:4
Mitigation: Zero data retention is architectural, not policy. Stateless design means there is nothing to breach — no database, no session storage, no logging of inputs. The /about page (roadmap item) will make the architecture explicit for enterprise buyers.
Residual Risk: 4 (reduced from 12)Owner: Product Manager
RISK-005MEDIUMPRODUCT / SCOPE

Users Treat Reliability Score as a Compliance Certificate Rather Than a Planning Tool

9
L:3 × I:3
Mitigation: Three distinct honesty markers in the UI: grounding disclosure, self-assessment disclaimer, AI-generated label. All exports include a liability disclaimer. The /about page will explain what TrueCase is and explicitly what it is not.
Residual Risk: 4 (reduced from 9)Owner: Product Design

OKRs & Success Metrics

Objective

Build the first governance-adjusted business case tool for AI investments — making regulatory maturity visible as a financial variable, not an afterthought, with zero data retention and fully grounded AI output

Key Results

Ship a 6-element governance gate backed by a verified KB with UK regulatory citations and weighted reliability scoring

100%

Target: All 6 elements live

Implement a reliability score (0–100%) that adjusts financial projections and is always shown alongside the headline figure at identical visual weight

100%

Target: Always co-equal display

Ground all Claude-generated narrative in governance-kb.json, benchmarks.json, and sdg-kb.json — no training data fallback

100%

Target: 100% KB-sourced claims

Deploy to production on Vercel with zero data retention architecture validated

100%

Target: Live public URL

Success Metrics

MetricTargetAchievedStatus
Governance KB elements6 elements verified6 — all with regulatory anchors + deductionsAchieved
Reliability score formula100% traceable deductions100% — all deductions from KBAchieved
Data retention0 bytes per session0 — stateless architectureAchieved
Production deploymentLive public URLtruecase-seven.vercel.appAchieved

Learnings & Reflections

What Went Well

  • The reliability score as the central product metaphor was the correct decision from day one. It gives both the finance team and the compliance team a single shared number — and the gap between headline and adjusted gain is more valuable than either projection alone.
  • Grounding the Claude system prompt explicitly in the KB — rather than relying on Claude's general regulatory knowledge — was the right call for a governance tool. Verifiability is the product's core promise. A hallucinated GDPR citation undermines everything.
  • Building stateless from the start, rather than adding privacy controls later, meant every architectural decision reinforced trust. There is no database to breach because there is no database. That is a stronger guarantee than any encryption policy.
  • The rule-based summary after Q1–Q4 (identifying the two most critical governance elements for the use case without an API call) is the highest-leverage component in the product. Four dropdowns → sector-specific plain English governance readiness assessment, instantly, at zero cost.

Challenges Faced

  • The Q1–Q4 → plain English summary required mapping every combination of four inputs to meaningful, sector-specific text. The edge cases (e.g. 'automation' × 'public sector' × 'unknown scale' × 'healthcare') multiplied faster than expected and required careful KB reference to avoid vague output.
  • Enforcing the boundary between what Claude's training data knows and what the KB specifies required iterative system prompt engineering. 'Use KB entries for all factual claims' is correct but insufficient — the prompt needs explicit negative examples to prevent confident-sounding additions.
  • The @react-pdf/renderer server-side PDF route is the most brittle component. React PDF requires hex values for colours (an explicit exception to the no-hardcoded-values rule), and layout behaviour diverges from the browser in ways that require a parallel styling pass.

What I'd Do Differently

  • Build the SDG knowledge base (sdg-kb.json) before the narrative section prompt, not alongside it. The Q1 × Q4 mapping complexity adds significant scope if discovered mid-prompt-engineering and requires KB-first thinking to structure correctly.
  • Ship the HTML export before the PDF export. HTML is simpler, gets the product live-demonstrable faster, and validates the narrative structure before investing in PDF layout. PDF can follow once the content model is proven.
  • Draft the /about page as part of Phase 1, not a roadmap item. The methodology explanation is part of the product's trust architecture — enterprise buyers and founders need to understand what TrueCase does and does not claim before they input sensitive financial data.

"The governance gap between headline ROI and reliability-adjusted ROI is not a failure state. It is the product. Every design decision — the co-equal card sizes, the self-assessment disclaimer, the KB-constrained Claude prompt, the stateless architecture — exists to make that gap credible enough to act on. A business case tool that always shows optimistic numbers is worse than no tool at all."

PM Artefacts

Written before any code. Every project ships with a full PM artefact set.

PRD — TrueCase
Model Decisions — TrueCase
Ethics Framework — TrueCase

Let's Connect

I am actively seeking Junior AI PM / Technical PM roles at companies building AI-powered products in media, events, e-commerce, or consumer applications. Let's connect if you're hiring or want to discuss AI product strategy.

© 2025 Ogbebor Osaheni. Built with Next.js, React, and Tailwind CSS.