WorkAboutLifeContact
Modguard.ai · Case Study 02 · LLM Trust & Safety
Designing trust into LLMs for healthcare and defense.
Around 100 user interviews, three core patterns shipped, and a product acquired with the patterns embedded — across two regulated verticals where verifying LLM output had to fit inside an expert workflow.
Client
Modguard.ai
Role
Design Advisor
Timeline
9 months
Verticals
Healthcare + Defense
Stage
Early → Acquired
Year
2023-2024
Where Modguard played
Two verticals, one product, a shared trust problem.
Vertical 01
Healthcare
Primary user: clinicians reviewing AI-generated patient summaries
High cost of error, time-bound between visits
Compliance: HIPAA, medical liability
The product
Modguard - LLM trust layer
StageEarly-stage AI/ML startup
MandateLLM solutions for regulated industries
My roleDesign Advisor, ~9 months
OutcomeAcquired with patterns intact
Vertical 02
Defense
Primary user: intel analysts validating LLM-synthesized briefs
Deadline-bound, fabricated-source risk
Compliance: classification, source provenance
Two domains, a shared trust problem: helping experts act on imperfect output without slowing them down.
The brief I reformulated
The brief I was given, and the brief I reformulated.
· The original brief
Make the LLM feel trustworthy.
"Reduce hallucinations. Add citations. Make the model sound more confident."
—Treats trust as an output-quality property
—Aims to make the model the source of confidence
—Underspecified: 2023 LLMs will hallucinate
Reframed
· The reframed thesis
Make verification cheap enough to absorb imperfect output.
Trust is a workflow property, not an output property.
1
Avoid making the model sound confident — that combines opacity with risk.2
Make the evidence easier to inspect, not the prose easier to accept.3
Treat the user as an expert who needs speed, not a novice who needs reassurance.The reformulation opened the design space. With trust framed as a workflow problem, the question shifted from "how do we improve the model?" to "how do we make verification fast enough that imperfect output becomes manageable?"
Two users
Two users. Both experts. Both time-bound.
The shared structure: domain experts under time pressure who pay a high cost when wrong. Different chrome, same workflow shape. The patterns we shipped had to work for both.
Three insights from research
~40 clinicians and ~60 analysts later, three findings shaped every pattern we shipped.
Design principle
Make uncertainty legible.
"
The model knows where its evidence is thin. The user usually doesn't.
→ The corollary
Verification must be cheap.
The cost to check has to approach zero, because experts will not trade speed for ceremony. If verification takes longer than skipping it, it won't happen.
The contested tradeoff
Speed vs. warning density: flag every uncertain claim, or only the most consequential? Flag too few → over-trust. Flag too many → alarm fatigue.
Where we landed: flag only the most consequential, with the recognition that this call would need to be revisited if a real adverse event surfaced. A decision made deliberately, with the tradeoffs documented.
Three patterns shipped
Hallucination flagging, citation visibility, guardrail indicators. Each addresses a different failure mode of LLM output, designed to fit inside an expert workflow.
What the work actually moved
Acquisition
Acquired
Modguard was acquired with the trust patterns intact in the product surface.
Verification time
~10xfaster
Per-citation verification time, from minutes-per-claim to seconds-per-claim in the analyst workflow.
Trust metric
CTR
Citation click-through rate adopted as a leading trust metric, replacing post-hoc trust surveys.
// the lesson I'm taking with me
In regulated industries, deferring well is more valuable than answering well. Asymmetric costs require asymmetric design.
// what generalizes
Trust is a workflow property.
Graded flagging, hover-to-verify, persistent competence boundaries — the patterns travel from clinical chart review to intel briefing to any expert-facing LLM surface.
Legal reviewFinancial reportingCompliance audit
Designed by Rebecka Raj · Modguard.ai · 2023-2024
The patterns shipped before the acquisition and survived it. They remain the spine of the trust surface in the acquiring company's regulated-industry product line.
Back to Projects