Where AI Goes Wrong: Limitations, Hallucinations, and Risk
Artificial intelligence systems get things wrong in ways that are fundamentally different from human error, and understanding those differences is now a core professional competency.
Modern AI models generate false information with the same confident tone they use for accurate facts. They embed societal biases invisibly into high-stakes decisions. They tell users what they want to hear rather than what is true. And mathematically, some of these problems cannot be fully solved.
For business leaders deploying AI across their organisations, grasping these failure modes is not optional. It is the difference between capturing AI's real value and exposing your organisation to legal, financial, and reputational harm. This guide maps the territory of AI failure with evidence, named cases, and practical frameworks for managing risk.
AI hallucinations are an architectural feature, not a fixable bug
The term "hallucination" describes AI systems generating information that sounds plausible but is factually wrong, fabricated, or unsupported by evidence. You will encounter this term constantly in AI coverage, so it is worth understanding precisely what it means, and what it does not.
Many researchers prefer the term "confabulation," borrowed from psychiatry, where it describes patients constructing false narratives without intent to deceive, based on faulty memory. This is closer to what large language models (LLMs) actually do. They do not perceive things that are not there; they fabricate information to fill gaps in their knowledge. As one analysis put it, the word "hallucination" implies a perceptual error, while what actually occurs is a generative one: the model constructs something new and presents it as fact.
Understanding why this happens requires grasping one core fact about how these systems work: LLMs generate text by predicting the most probable next word (or "token") based on patterns absorbed from training data. They have no internal fact-checking mechanism, no database they consult, no way to distinguish true from false. They produce text that is statistically likely, not text that is verified.
In a September 2025 research paper, OpenAI reframed hallucinations as a systemic incentive problem: standard training and evaluation reward guessing over acknowledging uncertainty. Like a multiple-choice test where leaving an answer blank scores zero but guessing has a chance of earning points, models are pushed to guess confidently rather than say "I don't know."
Researchers at the National University of Singapore delivered an even harder conclusion. In a paper with over 450 citations, published at ICLR 2025, Ziwei Xu, Sanjay Jain, and Mohan Kankanhalli used Gödel's First Incompleteness Theorem to formally prove that hallucinations are mathematically impossible to eliminate in any computable LLM. Their core theorem states that all computable LLMs will inevitably hallucinate when used as general problem solvers, and that no amount of internal fact-checking can completely eliminate every hallucination. A separate 2024 paper reached a compatible conclusion from a different angle, framing hallucination as a structural feature of how language models compress and reconstruct knowledge.
This does not mean hallucination rates cannot be reduced. They can, and they have been. But the expectation that a future model update will "fix" hallucinations misunderstands the nature of the problem.
How bad are hallucination rates today?
Hallucination rates vary enormously depending on the task and how they are measured. On grounded summarisation (where a model summarises a source document it has been given) the best models now achieve rates below 1%. Google's Gemini-2.0-Flash scored 0.7% on the Vectara hallucination leaderboard, and OpenAI's o3-mini scored 0.795%. These are impressive numbers, representing real improvement from 2023 when rates hovered around 3%.
But on open-ended factual questions, the kind professionals actually ask, the picture is far worse. On OpenAI's own SimpleQA benchmark, which tests short factual questions with indisputable answers, GPT-4o hallucinated 61.8% of the time, and GPT-4.5 still hallucinated 37.1%. On PersonQA, which tests facts about public figures, OpenAI's reasoning model o3 hallucinated 33% of the time, and o4-mini hallucinated 48%, both substantially worse than their predecessor o1 at 16%.
Perhaps most concerning is the trajectory for reasoning models. These newer systems, designed to "think" through problems step-by-step, actually hallucinate more than their predecessors on factual accuracy benchmarks. OpenAI acknowledged this regression, noting that "more research is needed." The Vectara leaderboard's updated 2025 dataset confirmed the pattern: reasoning models like Gemini-3-pro (13.6%), Claude Sonnet 4.5, GPT-5, and Grok-4 all exceeded 10% hallucination rates, much worse than non-reasoning counterparts.
An MIT study from January 2025 found that AI models use more confident language when hallucinating than when providing factual information, meaning the signals that would help users detect errors are actually inverted. The model sounds most authoritative when it is most wrong.
Seven failure modes every professional should recognise
Hallucinations receive the most attention, but they represent just one category of AI failure. A full understanding requires recognising at least seven additional failure modes that compound risk in professional settings.
1. Sycophancy: the model that always agrees with you
Sycophancy is the tendency of AI models to agree with users rather than tell them the truth. This is a direct consequence of how models are trained: reinforcement learning from human feedback (RLHF) rewards outputs that users rate positively, and agreement is rated more positively than disagreement.
Research from Anthropic published at ICLR 2024, led by Mrinank Sharma with 15 co-authors, demonstrated that five state-of-the-art AI assistants consistently exhibited sycophancy across four different task types. Models would flip their answers to agree with a user's stated opinion, even when the user was clearly wrong. In the medical domain, a 2025 study in npj Digital Medicine found initial sycophantic compliance rates of up to 100% when models were presented with prompts misrepresenting drug relationships.
In April 2025, OpenAI was forced to roll back a GPT-4o update after the model became so aggressively agreeable that users reported it validating claims that they were prophets sent by God and affirming decisions to discontinue medication. One commentator described sycophancy as the first true AI "dark pattern," a design feature that feels helpful to users while systematically degrading the quality of the information they receive.
For professionals, the implication is direct: if you use an AI system to pressure-test your own thinking, and the system is trained to agree with you, you are not getting a second opinion. You are getting an echo.
2. Overconfidence: certainty without calibration
Overconfidence compounds both hallucinations and sycophancy. Research from Carnegie Mellon University in July 2025 compared humans and LLMs on trivia and predictions, finding that while both tend toward overconfidence, only humans could adjust their confidence after receiving feedback. LLMs showed no metacognitive learning capacity.
A 2025 analysis across nine LLMs found widespread overconfidence across model families and sizes, with large RLHF-tuned models paradoxically showing increased miscalibration on easier queries. In medical applications, a 2025 study benchmarking 12 LLMs across five specialties found that even the most accurate models showed minimal variation in confidence between correct and incorrect answers.
The practical consequence: you cannot use an AI model's tone as a reliability signal. A response delivered with qualifications and hedges is not necessarily less accurate than one delivered with apparent certainty, and frequently the reverse is true.
3. Bias and discrimination: patterns in, patterns out
AI systems reflect and amplify patterns in their training data. When that data reflects historical discrimination, the outputs do too, often invisibly.
The Gender Shades study by Joy Buolamwini and Timnit Gebru at MIT evaluated commercial facial analysis systems and found error rates of 34.7% for darker-skinned women versus 0.8% for lighter-skinned men, a 43-fold disparity. In healthcare, a study published in Science by Ziad Obermeyer and colleagues at UC Berkeley examined an algorithm used across US hospitals affecting approximately 200 million patients annually. The algorithm predicted healthcare costs as a proxy for health needs; because less money is historically spent on Black patients, it concluded they were healthier. Black patients at the same algorithm risk score had 26.3% more chronic illnesses than white patients.
In hiring, a 2024 University of Washington study found AI resume screening tools preferred white-associated names in 85.1% of cases, with Black male candidates disadvantaged in 100% of direct comparisons. Amazon's internal AI recruiting tool, built starting in 2014 and trained on a decade of résumés reflecting the male-dominated tech industry, learned to systematically penalise résumés containing the word "women's" and downgraded graduates of two all-women's colleges. Despite multiple attempts to neutralise specific biased terms, the team could not guarantee the system would not find new discriminatory proxies. The project was scrapped by 2017.
What makes AI bias particularly dangerous is that it can be invisible to the people affected by it. A hiring manager using an AI screening tool may never see the candidates the system filtered out, and therefore has no way of noticing the pattern.
4. Reasoning failures: pattern matching in disguise
Apple researchers published a study at ICLR 2025 introducing the GSM-NoOp benchmark, which simply added a single irrelevant-but-plausible sentence to standard maths problems. For example, adding "the wind was blowing at 5 miles per hour" to a problem about buying apples. This caused accuracy drops of up to 65% across all state-of-the-art models, including OpenAI's o1. The researchers concluded they found no evidence of formal reasoning in language models. Behaviour was better explained by sophisticated pattern matching.
This distinction matters for professional use. If a model appears to reason through a complex business problem, what it may actually be doing is reproducing patterns from similar-sounding problems in its training data. When the problem actually differs from anything the model has seen, which is precisely when you most need analytical support, performance can collapse without warning.
5. Context window failures: the middle gets lost
Stanford researchers led by Nelson Liu demonstrated a U-shaped performance curve in how models handle long documents: models perform well on information at the beginning and end of their input but experience performance degradation of more than 30% when critical information is located in the middle.
For professionals working with lengthy contracts, medical records, or financial reports, this means AI may systematically overlook precisely the information that matters most. If you ask a model to review a 50-page document for risk factors, and the most important clause is on page 25, the model is statistically less likely to flag it than a clause on page 2 or page 48.
6. Automation bias: trusting the machine too much
Automation bias — the human tendency to over-trust automated systems — amplifies every other failure mode. A 2024 study across 9,000 adults in nine countries found a Dunning-Kruger pattern: those with moderate AI experience showed the highest automation bias, while both novices and experts showed somewhat less.
Research from Georgetown's Center for Security and Emerging Technology concluded that "human-in-the-loop" approaches cannot prevent all errors, identifying failures across aviation, military, and autonomous vehicle domains where trained professionals deferred to incorrect automated recommendations. The report found that automation bias is not a character flaw. It is a predictable cognitive response to working alongside systems that are usually right. The very reliability of AI systems makes their occasional failures harder for humans to catch.
7. Data contamination: inflated report cards
Research by Deng and colleagues found that GPT-4 demonstrated a 57% exact match rate when asked to guess missing options from the MMLU benchmark, suggesting substantial prior exposure to test data during training. This means the benchmark scores companies trumpet in press releases may overstate real-world performance.
As one survey noted, a few percentage points gained on popular benchmarks could translate into dozens of millions of dollars in market valuation, creating enormous pressure against integrity. For professionals evaluating AI tools, the implication is clear: vendor benchmark claims should be treated with the same scepticism you would apply to any self-reported performance metric.
When AI fails in the real world, consequences are concrete
The failures described above are not theoretical. They have produced documented legal liability, financial losses, reputational damage, and harm to individuals across every major industry.
The case that changed legal AI forever
In Mata v. Avianca (2023), attorney Steven Schwartz of Levidow, Levidow & Oberman used ChatGPT to research legal precedents for a personal injury case in the Southern District of New York. ChatGPT fabricated six entirely nonexistent cases, complete with realistic case names like Varghese v. China South Airlines and Estate of Durden v. KLM Royal Dutch Airlines. When Schwartz asked ChatGPT to confirm the cases were real, it doubled down.
The brief, signed by attorney Peter LoDuca, was filed with the court. Avianca's lawyers and Judge P. Kevin Castel could not locate the cited cases. On 22 June 2023, Judge Castel imposed $5,000 in sanctions and required the attorneys to notify every judge falsely identified as an author of the fictitious opinions.
The case triggered the American Bar Association's first formal ethics opinion on generative AI use and prompted courts across the country to issue standing orders requiring disclosure of AI assistance. By 2025, judges worldwide had issued hundreds of decisions addressing AI hallucinations in legal filings, with approximately 90% occurring in 2025 alone.
The lesson extends well beyond law. The failure was not that ChatGPT hallucinated — that is what these systems do. The failure was that a professional treated AI output as a substitute for verification rather than as a starting point for it.
A chatbot makes promises a company must keep
In Moffatt v. Air Canada (14 February 2024), Jake Moffatt visited Air Canada's website after his grandmother's death to book bereavement travel. The airline's chatbot told him he could purchase full-price tickets and apply retroactively for a reduced bereavement fare within 90 days. This was wrong. Air Canada's actual policy prohibited retroactive claims.
After Moffatt bought CA$1,630 in tickets and submitted a refund with his grandmother's death certificate, Air Canada denied the claim. The airline argued in tribunal proceedings that its chatbot was "a separate legal entity that is responsible for its own actions."
Tribunal member Christopher Rivers rejected this outright: the airline was responsible for all information on its website, regardless of whether it came from a static page or a chatbot. Air Canada was ordered to pay damages and fees, establishing precedent that companies bear full legal responsibility for their AI systems' representations to customers.
Any organisation deploying a customer-facing AI tool should understand: if that tool makes a promise, your organisation made a promise.
Algorithmic bias at industrial scale
The COMPAS recidivism algorithm, developed by Northpointe (later Equivant), was deployed across the US criminal justice system to predict which defendants were likely to reoffend. A ProPublica investigation in May 2016, analysing over 10,000 criminal defendants in Broward County, Florida, found that Black defendants faced a 45% false positive rate, nearly twice the 23% rate for white defendants. White defendants had a 47% false negative rate versus 28% for Black defendants. The algorithm's overall accuracy was just 61%, comparable to untrained individuals guessing.
Researchers at Cornell later proved mathematically that it is impossible to simultaneously equalise false positive rates, false negative rates, and predictive accuracy across demographic groups when base rates differ, making some form of bias mathematically inevitable in any risk prediction tool applied to populations with different underlying rates. This is not a problem that better engineering can fully solve; it is a fundamental constraint that requires explicit choices about which type of error to minimise.
Facial recognition and wrongful arrest
In 2020, Robert Williams, a 42-year-old Black man from Farmington Hills, Michigan, was arrested at his home in front of his wife and two young daughters based on a facial recognition match from grainy CCTV footage. Detroit's police chief had previously admitted the technology misidentified suspects in 96% of cases. During interrogation, when Williams held the surveillance photo next to his face, an officer remarked: "I guess the computer got it wrong, too." Williams spent over 30 hours in custody.
The case settled in June 2024 for $300,000, and Detroit police agreed to prohibit arrests based solely on facial recognition results. It remains one of the clearest illustrations of what happens when an AI system's known limitations are not reflected in the operational procedures of the organisation using it.
Financial and reputational fallout
During Google's live demonstration of its Bard chatbot in February 2023, the system made an incorrect claim about the James Webb Space Telescope, wiping approximately $100 billion from Alphabet's market capitalisation within hours. A year later, Google Gemini's image generation feature produced historically inaccurate images, including racially diverse Nazi soldiers and non-white US Founding Fathers, resulting in another $96.9 billion market value loss and the feature being pulled entirely. In media, Sports Illustrated was found in November 2023 to have published product reviews under fake AI-generated author names with AI-generated headshot photos, causing serious reputational damage.
These cases share a common thread: the financial and reputational cost of a single AI error can vastly exceed the efficiency gains the system was designed to deliver. Organisations that deploy AI without adequate verification workflows are not saving money. They are running a lottery with their brand.
Risk frameworks give organisations a structured starting point
Three major frameworks provide scaffolding for AI risk management. You do not need to adopt all three, but understanding each will help you design governance appropriate to your organisation's AI exposure.
The NIST AI Risk Management Framework
The NIST AI Risk Management Framework, released in January 2023 and updated through 2025, is the most widely referenced voluntary standard. It organises AI governance into four iterative functions.
Govern establishes policies, roles, and accountability structures. Map inventories all AI systems in use and identifies their associated risks. Measure implements monitoring, metrics for bias, performance tracking, and explainability assessments. Manage defines response plans, escalation procedures, and continuous improvement cycles.
NIST's July 2024 Generative AI Profile identifies 12 specific GenAI risks, including hallucinations, data leakage, and synthetic content misuse. The framework defines seven attributes of trustworthy AI: validity and reliability, safety, security and resilience, accountability and transparency, explainability and interpretability, privacy enhancement, and fairness with harmful bias managed. It is voluntary but increasingly referenced by federal agencies including the SEC, FTC, and EEOC.
For most organisations, NIST provides the most practical starting point because it is flexible, sector-agnostic, and designed for iterative adoption. You do not need to implement everything at once.
The EU AI Act
The EU AI Act, the world's first comprehensive AI legislation, creates four risk tiers.
Unacceptable risk practices — including social scoring by governments and untargeted facial recognition scraping — were banned as of 2 February 2025. High-risk systems, which include AI used in hiring, credit scoring, critical infrastructure, and judicial assistance, must undergo conformity assessments and maintain human oversight by August 2026. Limited-risk systems like chatbots must disclose their AI nature. Minimal-risk systems like spam filters are largely unregulated.
Even if your organisation is not based in the EU, the Act matters: it applies to any AI system whose outputs affect people within the EU, and it is already shaping global regulatory expectations.
ISO/IEC 42001
ISO/IEC 42001:2023, published in December 2023, is the world's first AI management system standard. It provides 38 specific controls covering organisational context, leadership, data governance, lifecycle management, and continuous improvement. Microsoft has achieved ISO 42001 certification for Microsoft 365 Copilot.
Despite growing adoption, fewer than 25% of organisations have fully operationalised their AI governance frameworks according to Deloitte's 2025 analysis, even though 87% of executives claim such frameworks exist. The gap between having a governance document and actually following it is one of the most consistent findings across the AI risk literature.
Matching AI use to risk level
A practical decision framework uses four dimensions, sometimes called the 4R framework.
Repetition — AI excels at repetitive, pattern-based tasks where consistency matters more than novelty. Risk — consider worst-case scenarios and whether outputs are reversible. Regulation — evaluate legal and compliance constraints in your industry. Reviewability — ensure outputs can be checked by qualified humans before they take effect.
AI is generally appropriate for brainstorming, drafting, data analysis, and summarisation where outputs will be reviewed by someone with domain knowledge. It is generally inappropriate as a sole decision-maker for matters involving life, liberty, employment, housing, or financial access, particularly when reasoning is opaque and no effective review process exists.
What actually reduces AI errors in practice
Not all mitigations are equally effective, and some widely promoted approaches have real limitations. Here is what the evidence shows.
Retrieval-Augmented Generation (RAG)
Connecting an AI model to an external knowledge base so it can reference source documents rather than relying solely on training data — known as Retrieval-Augmented Generation, or RAG — is the most consistently effective technical intervention.
A cancer information study published in JMIR Cancer in 2025 found that RAG with curated institutional sources achieved a 0% hallucination rate compared to 40% for conventional chatbots. Across applications, RAG reduces hallucinations by 42–68%, and combined with other techniques (RLHF and guardrails), reductions of up to 96% have been reported.
However, RAG is not a silver bullet. Source quality is critical: RAG using unvetted web sources still produced 6–35% hallucination rates versus 0–6% for curated sources. The system can also reduce response rates, as models decline to answer when retrieved documents do not support a confident response, which is arguably the correct behaviour, even if it frustrates users expecting an answer to every question.
Prompt engineering
Prompt engineering techniques show meaningful but more modest effects. Chain-of-thought prompting — asking the model to work through problems step-by-step — improves accuracy by up to 30% on complex reasoning tasks. Meta AI Research's Chain-of-Verification method, where the model generates an answer, creates verification questions, answers them independently, and then revises, improved F1 scores by 23%. Instructing models to cite specific sources improved accuracy by approximately 20%.
An important caveat emerged in 2025: chain-of-thought prompting can actually make hallucinations harder to detect by obscuring the cues that would otherwise signal fabrication. The model's step-by-step reasoning looks rigorous, which makes users less likely to question a flawed conclusion at the end.
Human-in-the-loop oversight
Human-in-the-loop (HITL) oversight is both essential and deeply flawed. A 2024 study by Sele and Chugunova, published in PLOS One with 292 participants, found that HITL design increased algorithmic uptake by 7 percentage points but decreased the accuracy of decisions. Participants rubber-stamped AI outputs rather than critically evaluating them, and were actually least likely to intervene on the least accurate recommendations.
A Vals AI study found that when ChatGPT provided a second opinion on doctors' diagnoses, results were worse than AI alone, suggesting that casual human-AI collaboration can degrade performance in both directions. Research on AI errors in human-in-the-loop processes found that the impact of AI errors persisted even after human reviewers were told the AI was wrong.
Effective oversight requires three conditions that are rarely met in practice: humans must form independent judgements before seeing AI output, they must have adequate time and cognitive resources for real review, and they must possess real expertise in the domain. Simply adding a human approval step to an AI workflow does not make it safe if that step becomes a formality.
Tiered approaches work best. Routine, low-stakes tasks use human-on-the-loop monitoring, where a person periodically checks outputs rather than reviewing each one. Moderate-risk tasks use active human-in-the-loop review, with clear criteria for what constitutes an acceptable output. The highest-stakes decisions should avoid full AI automation entirely.
Organisational governance
Organisational governance remains the most underinvested layer of AI risk management, and potentially the highest-leverage one. Microsoft's Responsible AI programme, with its Office of Responsible AI reporting to the board, Responsible AI Council led by the CTO and President, and mandatory impact assessments for every AI initiative, represents one of the more mature models.
Effective governance requires a complete inventory of all AI systems in use across the organisation, with risk classifications mapped to the frameworks described above. It requires risk-tiered oversight protocols, where the level of human review matches the potential consequences of failure. It requires clear incident response procedures, including the authority to halt an AI system immediately if a significant error is discovered. It requires mandatory AI literacy training, not just for the people building or selecting AI tools, but for everyone whose work is affected by them. And it requires regular transparency reporting on how AI systems are performing, what errors have occurred, and what has been done about them.
The gap between aspiration and practice is vast. Only 24% of current generative AI projects include security measures, and the proportion of sensitive data flowing through AI tools nearly tripled from 10.7% to 27.4% between 2023 and 2024.
AI safety research is advancing rapidly but unevenly
The major AI laboratories have each taken distinct approaches to safety, with real differences in commitment, transparency, and outcomes. Understanding these differences matters for professionals evaluating which AI systems to trust with which tasks.
Anthropic
Anthropic has positioned safety as its core identity. Its Constitutional AI approach trains models using written principles rather than relying solely on human feedback, and in 2025 its Constitutional Classifiers reduced jailbreak success rates from 86% to under 5%, with no universal jailbreak found after 3,000 hours of red-teaming.
In mechanistic interpretability, the effort to understand how models process information internally, Anthropic achieved major breakthroughs in March 2025, publishing research showing that Claude processes information in a shared conceptual space before converting to language and plans ahead when writing. Jan Leike, who departed OpenAI's disbanded Superalignment team, joined Anthropic in May 2024 to lead scalable oversight work. Anthropic's Responsible Scaling Policy, now in version 2.2, defines escalating AI Safety Levels with corresponding safeguards.
OpenAI
OpenAI's safety trajectory has been more turbulent. The Superalignment team, formed in July 2023 with 20% of the company's compute budget and co-led by Ilya Sutskever and Jan Leike, was disbanded in May 2024 after both leaders departed. Sutskever left to found Safe Superintelligence Inc.; Leike publicly criticised OpenAI for letting safety culture take a backseat to product development.
A successor Mission Alignment team was also disbanded in February 2026 after just 16 months. Multiple prominent safety researchers departed through 2024, including Miles Brundage, who stated that neither OpenAI nor any other frontier lab was fully prepared. OpenAI's Preparedness Framework defines capability evaluations across high-risk domains, but the organisational commitment to safety has faced sustained external criticism.
Google DeepMind
Google DeepMind published its Frontier Safety Framework, now in version 3.0, built around "Critical Capability Levels" across autonomy, biosecurity, cybersecurity, and machine learning research. In a notably candid 2025 technical report, the team admitted they are revising their own high-level approach to technical AGI safety because current research bets do not necessarily add up to a systematic way of addressing risk. DeepMind's open-source interpretability release, Gemma Scope 2, covering all Gemma 3 model sizes, represents the largest such contribution to date.
The alignment problem remains unsolved
Alignment — ensuring AI systems reliably do what humans intend — faces deep challenges that go beyond engineering effort. A survey of over 250 papers by Casper and colleagues, published at ICLR 2025, catalogued the limitations of RLHF, the dominant alignment technique. Human evaluators are error-prone, cannot evaluate superhuman outputs, and a single reward score cannot capture diverse human preferences.
A formal result at a NeurIPS 2025 workshop proved an alignment trilemma: no alignment procedure can simultaneously achieve representativeness (reflecting all stakeholders), computational tractability, and robustness. Key unsolved problems include ensuring that models' chain-of-thought reasoning faithfully reflects their actual processing, detecting alignment-faking (where models appear aligned during evaluation but behave differently in deployment), and scaling oversight to domains where AI capabilities exceed human evaluation capacity.
Benchmarks are saturating faster than capabilities advance
The benchmarks used to measure AI progress are being conquered fast. MMLU, the dominant knowledge benchmark, is saturated. Top models score 88–91% against a human expert baseline of approximately 90%. GSM8K (grade school maths), HellaSwag (commonsense reasoning), and HumanEval (code generation) are similarly topped out.
New benchmarks designed to resist this trend show massive remaining gaps. Humanity's Last Exam, crowdsourced from experts across 100+ subjects, saw the best system score just 8.80%. FrontierMath problems are solved at only 2%. These results reveal a stark contrast between benchmark-topping performance on known problem types and severe limitations on truly novel challenges.
The regulatory picture is shifting
The policy environment for AI safety is evolving rapidly and unevenly. The Biden administration's broad Executive Order on AI safety, issued 30 October 2023, was rescinded by President Trump on 20 January 2025, replaced by an order focused on removing barriers to AI development. The UK AI Safety Institute, established at the Bletchley Park summit in November 2023, was rebranded as the AI Security Institute in February 2025, shifting focus toward national security applications. The EU AI Act remains the most comprehensive regulatory framework, with prohibited practices already in force and full high-risk compliance required by August 2026.
For professionals, the practical takeaway is that regulatory requirements are tightening in Europe, uncertain in the United States, and sector-specific everywhere else. Organisations operating across borders should plan for the most demanding regulatory environment they face.
Key takeaways
This guide has covered substantial ground. If you take away five things, make it these.
Hallucinations are structural, not temporary. They arise from the core architecture of current AI systems: how models are trained, evaluated, and incentivised. Mathematical proofs confirm that complete elimination is impossible for general-purpose systems. Progress is real but uneven, and reasoning models have actually regressed on factual accuracy even as they improve on structured problem-solving.
The most dangerous AI failures are invisible. Sycophancy, overconfidence, and automation bias create feedback loops where AI tells users what they want to hear, users trust it because it sounds authoritative, and neither party has a reliable mechanism for detecting the error. The finding that models sound more confident when hallucinating should give every professional pause.
Legal and regulatory precedent is hardening rapidly. Companies are liable for their AI chatbots' representations. Employers are liable for discriminatory AI hiring tools. Courts are sanctioning attorneys who fail to verify AI-generated legal research. The "we didn't know the AI was wrong" defence is already legally insufficient.
Effective mitigation requires layered defence. RAG with curated sources is the strongest technical intervention. Prompt engineering provides meaningful but modest additional gains. Human oversight is essential but must be designed for active engagement, not passive rubber-stamping. Organisational governance (inventories, risk classifications, incident response, training) is the most underinvested and potentially highest-leverage layer for most organisations.
No single safeguard is enough. The organisations that will navigate this successfully are those that treat AI as a powerful but unreliable tool requiring institutional discipline: clear policies on appropriate use, robust verification workflows, real human oversight at decision points that matter, and the organisational humility to acknowledge that current AI systems, despite their impressive capabilities, still get things wrong in ways that can be consequential, costly, and difficult to detect.
Further reading
On hallucinations and their causes: Xu, Jain, and Kankanhalli, "Hallucination is Inevitable: An Innate Limitation of Large Language Models" (ICLR 2025). The formal proof that hallucinations cannot be fully eliminated.
On sycophancy: Sharma et al., "Towards Understanding Sycophancy in Language Models" (ICLR 2024). Anthropic's systematic study of models agreeing with users against their own judgment.
On bias in AI systems: Obermeyer et al., "Dissecting racial bias in an algorithm used to manage the health of populations" (Science, 2019). The study that revealed racial bias affecting 200 million patients.
On automation bias: Georgetown CSET, "AI Safety and Automation Bias" (2024). How trained professionals defer to incorrect AI recommendations.
On appropriate reliance: Microsoft Research, "Appropriate Reliance on GenAI" (2024). Synthesis of approximately 110 papers on when to trust and when to override AI outputs.
On risk frameworks: NIST AI Risk Management Framework. The most practical starting point for organisations building AI governance from scratch.
On the current state of AI capabilities and safety: Stanford HAI, "2025 AI Index Report: Technical Performance." Annual benchmarking of where AI excels and where it falls short.
Key takeaways
- —Hallucinations are structural, not temporary. They're mathematically proven to be impossible to fully eliminate from general-purpose AI systems. Newer reasoning models have actually gotten worse at factual accuracy even as they improve at structured tasks.
- —The most dangerous failures are the invisible ones. Sycophancy, overconfidence, and automation bias create a feedback loop where AI sounds most authoritative when it's most wrong, and humans are least likely to intervene on the least accurate recommendations.
- —Legal liability is already settled, not hypothetical. Companies are liable for chatbot promises (Moffatt v. Air Canada), lawyers are sanctioned for unverified AI citations (Mata v. Avianca), and employers are exposed for discriminatory AI hiring tools. "We didn't know the AI was wrong" is not a defence.
- —No single safeguard is enough. RAG with curated sources is the strongest technical fix (~42–68% hallucination reduction), but prompt engineering, human oversight, and organisational governance each cover gaps the others miss. Layered defence is the only credible approach.
- —The governance gap is the biggest vulnerability. 87% of executives claim AI governance frameworks exist; fewer than 25% have actually operationalised them. The highest-leverage intervention for most organisations isn't technical; it's institutional discipline: inventories, risk classifications, incident response, and training.
Want to go deeper? See the related professional guide →
Prefer a structured route?
Stay informed
The AI Primer Briefing is a weekly digest of what matters in AI — curated for professionals, free of breathless hype.