A Scenario Playing Out in Boardrooms Across Hong Kong Right Now
A regional insurance firm's legal team has been using an AI research assistant for three months. The tool has been producing well-structured summaries, pulling case references, and saving hours of manual work. One morning, an associate presents a regulatory brief to the compliance committee that includes three case citations that do not exist. The AI generated them with complete confidence — accurate formatting, plausible case names, a convincing argument structure. Nobody caught it until a senior partner checked the source.
This is not a hypothetical. It is the most common form of AI failure in enterprise deployments today, and it has a name: hallucination. Understanding what AI hallucination is, why it happens, and how to systematically reduce its impact is not an IT question — it is a governance and risk management question that sits squarely in the CFO's and COO's domain.
What Is AI Hallucination — and Why the Term Matters Strategically
AI hallucination is the phenomenon where a large language model generates output that is factually incorrect, misleading, or entirely fabricated — presented with the same fluency and confidence as accurate information. The model does not "know" it is wrong. It does not flag uncertainty. It produces the most statistically probable sequence of words given its training, which sometimes results in plausible-sounding falsehoods that are indistinguishable from truth without expert verification.
The term matters strategically because it reframes how enterprise leaders should think about AI deployment. The question is not "Is this AI system accurate?" — because no current AI system is fully accurate across all inputs. The question is: "Under what conditions does this system hallucinate, how frequently, and what controls do we have to catch errors before they cause business, legal, or reputational harm?"
Gartner's March 2026 research on LLM observability predicts that by 2028, 50% of enterprise GenAI deployments will include formal LLM observability investment — up from 15% today. This trajectory reflects the industry's acknowledgement that hallucination is not a temporary bug being fixed in the next model version; it is a fundamental characteristic of current AI architecture that requires systematic management.
Why Do AI Systems Hallucinate — the Architecture Behind the Problem
Large language models generate text by predicting the most likely next token based on patterns learned from training data. They do not retrieve verified facts from a structured database, and they do not maintain a "knowledge map" of what they know versus what they are guessing. When asked a question that falls outside their training data, or when their training data was conflicting or incomplete, the model fills the gap with plausible-sounding content rather than reporting uncertainty.
Three structural conditions increase hallucination frequency in enterprise contexts. First, knowledge staleness: models trained on data up to a certain date cannot accurately answer questions about events, regulations, or market conditions that postdate their training. Second, specificity pressure: when users ask highly specific questions — exact regulatory article numbers, precise case citations, specific contract clause references — models are more likely to fabricate specifics than to admit they do not have the precise information. Third, context length: very long documents or complex multi-step reasoning chains increase error propagation, where a small incorrect assumption early in a chain of reasoning compunds into a materially wrong conclusion.
For Hong Kong enterprises, the practical implication is that hallucination risk is highest in exactly the use cases where AI is most attractive: legal and regulatory research, financial analysis requiring precise citations, compliance documentation, and complex client-facing communications. These are also the use cases where a single undetected error can create significant liability.
The Enterprise Risk Taxonomy: How to Classify Hallucination Exposure
Not all hallucinations carry equal risk. A structured risk taxonomy helps enterprise leaders prioritise control investment. Three tiers of hallucination exposure apply in most enterprise AI deployments.
Tier 1 — High-Stakes Factual Claims: outputs that will be presented to clients, regulators, or counterparties as factual assertions. Examples include regulatory compliance reports, legal research summaries, financial analysis for investment committees, and due diligence documentation. A hallucination in this tier can create direct legal liability and reputational damage. This tier requires mandatory human expert review before any output leaves the organisation.
Tier 2 — Internal Decision Support: outputs used by staff to inform internal decisions — market analysis, competitive intelligence, HR policy summaries, operational planning. Hallucinations here create decision-quality risk rather than direct external liability. This tier requires structured human review workflows and output confidence scoring where available.
Tier 3 — Process Automation and Routine Drafting: outputs that automate repetitive, low-stakes tasks — meeting summaries, email drafts, data formatting, initial document templates. Hallucinations here are typically caught by light human review before any action is taken. This tier can operate with periodic sampling and automated guardrails rather than full expert review.
Four Proven Technical Controls That Reduce Hallucination Risk
Enterprise technology teams have four primary technical controls for reducing hallucination frequency and impact. None eliminates hallucination entirely, but together they create a governance architecture that makes hallucination manageable.
Control 1 — Retrieval-Augmented Generation (RAG): Instead of relying solely on what the model was trained on, RAG systems retrieve relevant verified documents or data at query time and give the model explicit source material to work from. Hong Kong asset management firms using RAG-enabled systems report significant reductions in citation errors in equity research workflows, because the model is generating summaries of documents it has actually been given, rather than recalling information from training. Gartner recommends RAG as the primary mitigation for enterprise hallucination in knowledge-intensive workflows.
Control 2 — Output Guardrails and Confidence Scoring: Modern enterprise AI platforms can be configured with guardrails that flag low-confidence outputs, detect common hallucination patterns (e.g. URLs that don't exist, citation formats with implausible dates), and route uncertain responses to human review queues rather than auto-delivering them to users. Gartner's April 2026 guidance specifically recommends that general counsel assess AI guardrail architectures as part of AI risk management frameworks.
Control 3 — LLM Observability Infrastructure: LLM observability tools monitor model outputs in production, tracking which query types generate the highest error rates, which user groups encounter the most hallucinations, and whether error rates change as the underlying model updates. Gartner predicts 50% of enterprise GenAI deployments will adopt formal LLM observability by 2028 — organisations that implement this infrastructure now will have a two-year data advantage in calibrating their hallucination controls.
Control 4 — Prompt Engineering Standards: Many hallucinations are preventable through better prompt construction. Prompts that specify the format of the required response ("List only facts that appear in the provided document"), require the model to flag uncertainty ("If you are not certain of a specific date, say so explicitly"), and provide ground-truth context ("Answer only based on the following regulatory text: [text]") dramatically reduce hallucination rates. McKinsey's 2026 enterprise AI analysis found that organisations achieving top-quartile accuracy from AI systems invest systematically in prompt engineering training for all AI users, not just technical teams.
The Governance Framework: Building a Hallucination Risk Register
Technical controls alone are insufficient without an organisational governance framework. Enterprise leaders should build a hallucination risk register — a structured document that maps each AI use case to its hallucination risk tier, the technical controls in place, the human review process, and the escalation protocol if an undetected hallucination causes harm.
A hallucination risk register has four columns. Column one lists the use case (e.g. "regulatory research for compliance team"). Column two assigns the risk tier (1, 2, or 3). Column three lists the controls in place — RAG, guardrails, prompt standards, human review frequency. Column four defines the remediation protocol if a Tier 1 hallucination reaches a client or regulator — who owns the response, what the notification procedure is, and what the AI system suspension criteria are.
This framework serves a dual purpose. Internally, it creates clear accountability and ensures that every AI deployment has been reviewed against a consistent risk standard. Externally, it demonstrates to regulators — including Hong Kong's OPC and HKMA — that the organisation is managing AI risks with the same rigour it applies to other operational risk categories. In 2026, this is increasingly a regulatory expectation rather than best practice.
The Leadership Posture: Neither Paralysis Nor Denial
Enterprise leaders who understand AI hallucination tend to adopt one of two dysfunctional responses: paralysis (refusing to deploy AI in any high-stakes context until the technology is "perfect") or denial (treating hallucination as a minor inconvenience that careful users will catch). Neither response is strategically defensible.
The correct leadership posture is risk-proportionate deployment. Deploy AI in high-value, high-stakes contexts, but with the governance architecture — risk tiers, technical controls, review workflows, and a risk register — that makes the deployment auditable and correctable. The organisations that will lead in AI-enabled operations over the next three years are not the ones that waited for a perfect technology. They are the ones that built the governance infrastructure to manage an imperfect technology safely and at scale.
懂AI的冷,更懂你的難 — UD同行28年,讓科技成為有溫度的陪伴。 AI hallucination is not a reason to stop your AI programme. It is a reason to build one properly — with the controls, governance, and expertise that turn a powerful but imperfect technology into a reliable business asset.
Deploy AI with Confidence, Not Just Optimism
UD's AI Staff Solution is built with enterprise reliability controls — governed workflows, human review checkpoints, and the 28-year operational experience of a partner who has managed technology risk in Hong Kong enterprises across every market cycle. We'll walk you through every step: from use case risk assessment and control design to production deployment with the governance architecture your CFO and compliance team need to sign off.