Enterprise
AI Security
Threat Model
AI systems have expanded the enterprise attack surface beyond what traditional security frameworks were built to address. These are the nine critical threat vectors — with documented real-world incidents, risk impact analysis, and operational mitigation controls for each.
AI Has Expanded the Enterprise Attack Surface Beyond What Traditional Security Handles
// 9 attack vectors documented · OWASP LLM Top 10 mapped · NIST AI RMF aligned
CRITICAL: Prompt Injection (CVE-2025-53773, CVSS 9.6)
CRITICAL: API & Credential Theft (300K AI credentials in 2025 infostealers)
HIGH: Data Poisoning — Check Point names it “new zero-day” for AI systems
HIGH: Supply Chain (700+ orgs breached via single OAuth token — UNC6395, 2025)
HIGH: Excessive Autonomy — OWASP Agentic Top 10 published late 2025
MEDIUM: Model Inversion · Model Drift · Context Leakage · Compliance Violations
// Threat landscape is expanding faster than security frameworks can adapt.
// Continuous reassessment is mandatory — point-in-time audits are insufficient.
The AI security threat landscape of 2026 is defined by a convergence of traditional cybersecurity vectors and an entirely new class of attack that specifically exploits the probabilistic, data-dependent, and autonomous characteristics of AI systems. Attackers no longer need to hack into a server or exploit a code vulnerability to compromise an enterprise AI system — they can tamper with the data that trains it, inject instructions through the content it processes, steal the credentials that authenticate it, or simply wait for its self-supervised drift to produce harmful outputs without any active attack at all.
Check Point’s 2026 Tech Tsunami report calls prompt injection and data poisoning the “new zero-day” threats — attacks that blur the line between security vulnerability and misinformation. Microsoft’s RSAC 2026 analysis confirmed that AI is not just accelerating cyberattacks; it is upgrading them. The tempo, iteration speed, and precision of attacks have fundamentally changed even when the objectives — credential theft, financial gain, and espionage — remain constant.
The nine threat vectors mapped below each follow the same structure: Attack Vector (how the attack occurs), Risk Impact (what damage it causes), and Mitigation Strategy (the controls that address it). This is a practical operational reference, not a theoretical framework. Every documented incident cited is real and verified.
AI systems operating in regulated industries — financial services, healthcare, legal, insurance — face a compliance risk that does not require any external attacker: the system produces outputs that violate regulatory requirements on its own. An LLM-powered loan decision tool that produces discriminatory outputs violates the EU AI Act’s high-risk AI obligations and exposes the organisation to fines up to €35 million. A clinical AI that provides medical guidance beyond its authorised scope creates HIPAA exposure and professional liability.
The compliance failure is compounded by AI’s scale — a single non-compliant AI configuration can produce thousands of violating outputs before any human reviewer identifies the pattern. Automated pipelines that route AI outputs directly to customers or downstream systems accelerate the exposure window and the audit burden simultaneously.
The EU AI Act’s August 2026 enforcement date makes this threat operationally urgent: organisations that cannot demonstrate documentation, human oversight records, bias test results, and conformity assessments for high-risk AI systems face both fines and reputational damage. Compliance risk is no longer a legal department concern — it is an AI engineering and deployment obligation.
Data poisoning attacks inject malicious, false, or biased data into an AI system’s training pipeline or retrieval index — corrupting the model’s behaviour at the source rather than at the output layer. Unlike traditional software attacks, poisoning an AI does not require hacking into a server or exploiting a code bug — it only requires tampering with the data supply chain. Check Point’s 2026 Tech Tsunami report names data poisoning a “new zero-day” threat precisely because it subverts an organisation’s AI logic without touching its traditional IT infrastructure.
The scale required is smaller than intuition suggests. Research from Columbia, NYU, and Washington University demonstrated that as few as 50,000 fake articles added to a public training dataset were sufficient to corrupt medical LLMs — while another study found that very small quantities of poisoned data corrupted even the largest models. In 2025, successful poisoning attacks were carried out against RAG pipelines, MCP tool integrations, and synthetic data generation workflows. A single poisoned dataset can propagate across thousands of applications that depend on that model.
OWASP published a dedicated Top 10 for Agentic Applications by late 2025, confirming that autonomous AI agent risks now constitute a distinct security category. When AI agents are granted broad tool access, API permissions, and decision-making authority without corresponding oversight controls, they can take consequential actions far outside the scope their operators intended — deleting records, executing financial transactions, publishing content, or modifying infrastructure configurations.
The OpenClaw incident of 2026 — where a viral open-source AI agent with 135,000 GitHub stars created over 21,000 exposed enterprise instances — illustrated the scale of the risk. When employees connect autonomous agents to corporate systems like Slack, Google Workspace, or production databases, they create shadow AI with elevated privileges that traditional security tools cannot detect. The agent acts, at machine speed, on behalf of whoever deployed it — with no per-action human review.
Verizon’s 2025 Data Breach Investigations Report found that third-party involvement in breaches doubled year-over-year. The AI supply chain has dramatically expanded this surface: every third-party model, dataset, plugin, API integration, and library dependency is a potential vector for hidden vulnerabilities to enter enterprise AI systems before they are deployed. Unlike traditional software supply chain attacks, AI supply chain compromises can be invisible at the code level — a poisoned model checkpoint behaves identically to a clean one under normal operation but embeds backdoors that activate under specific trigger conditions.
The UNC6395 OAuth supply chain attack of August 2025 required no exploit. The attacker used stolen OAuth tokens from a trusted SaaS integration to access customer Salesforce environments across 700+ organisations — each connection looked legitimate because it came from a sanctioned SaaS app, not a compromised user account. This is the supply chain risk in its most operationally dangerous form: trust weaponised against the organisations that extended it.
Model drift is the gradual degradation of AI model performance as the statistical properties of production data diverge from those present in the training dataset. Unlike the other threats in this model, drift requires no active attacker — it is a passive failure mode inherent to any AI system operating in a dynamic world. Performance degradation is particularly dangerous because systems continue operating well enough until significant harm has already occurred.
Three drift variants require continuous monitoring: data drift (input distributions diverge from training), concept drift (the relationship between inputs and target outcomes changes over time), and upstream data drift (changes in data collection or processing alter incoming characteristics without any real-world change). A credit scoring model trained on pre-recession economic data encounters fundamentally different applicant profiles during an economic downturn — making decisions that appear algorithmically valid but are systematically miscalibrated to current conditions.
Model inversion attacks systematically query a deployed AI model and analyse its outputs to reconstruct information about its training data — recovering Personally Identifiable Information (PII), proprietary records, or confidential datasets that the model was trained on, without ever having access to those datasets directly. The attack requires only API access to the model — which may be publicly available.
When models memorise training data rather than learning patterns, information can leak via well-scoped queries never intended to retrieve it. This is especially severe for models fine-tuned on sensitive domain data — medical records, financial information, legal documents, internal communications. A healthcare organisation that fine-tunes an LLM on patient records and then deploys it via API has potentially created a recoverable store of protected health information accessible to anyone with an API key.
IBM’s 2026 X-Force Threat Intelligence Index found over 300,000 ChatGPT credentials discovered in infostealer malware in 2025. Stolen AI platform credentials pose risks far beyond another account entry point: attackers can siphon entire conversation histories filled with sensitive business information, access connected enterprise integrations, and use the legitimate account to make API calls that appear authorised. The Internet Archive’s catastrophic 2024 breach resulted from a GitLab authentication token hardcoded in 2022 — sitting dormant and unrotated for two years before exploitation.
AI API keys present unique risks because they are high-value, often long-lived, and frequently embedded in application code, environment variables, or CI/CD pipelines. Once compromised, an API key provides access not just to the model but to every integration that key has been granted permission to access — including RAG retrieval systems, tool call endpoints, and connected enterprise data stores.
Unauthorized tool invocation occurs when an AI agent calls a function, API endpoint, or tool it was not supposed to invoke for a given task — either because a prompt injection instructed it to, or because the model hallucinated an appropriate tool call and selected incorrectly from its available function set. The consequence is not incorrect text — it is an incorrect real-world action executed with the full permissions of the agent.
In 2025, the first major AI agent security crisis came from precisely this vector: OpenClaw agents connected to corporate Slack workspaces and Google Workspace began taking actions outside their intended scope when users connected them to production systems. Tool proliferation is the primary risk amplifier — an agent with 50 tools in its schema will mis-select far more often than one with 5. Every tool the agent can invoke is a surface for hallucinated invocations and adversarial injection.
Context leakage occurs when an AI system exposes information it should not — through its outputs, its reasoning trace, its system prompt, or through data accessible in its retrieval context. The EchoLeak vulnerability (CVE-2025-32711) in Microsoft 365 Copilot demonstrated the most dangerous form: a zero-click prompt injection embedded in a normal business document silently exfiltrated enterprise data without any user action. No code was executed. Copilot processed the malicious instruction as if it were a legitimate part of the document.
Context leakage is compounded by multi-tenant AI deployments where multiple users or organisations share a model instance without adequate context segmentation. When one user’s interaction can surface information loaded into the model’s context for another user’s session, the result is a cross-tenant data breach that is architecturally trivial to exploit and extremely difficult to detect after the fact.
“AI is not just accelerating cyberattacks — it is upgrading them. The objective has not changed: credential theft, financial gain, and espionage. What has changed is the tempo, the iteration speed, and the ability to test and refine at scale. Organizations building security around a static checklist will find their controls outdated before the next budget cycle.”
Microsoft Security Blog — Threat Actor Abuse of AI Accelerates From Tool to Cyberattack Surface · RSAC 2026All 9 Vectors — Quick Reference
Severity, OWASP mapping, and primary mitigation control for every documented threat.
| # | Threat Vector | Severity | OWASP / Reference | Primary Mitigation |
|---|---|---|---|---|
| 01 | Compliance & Regulatory Violations | MEDIUM | EU AI Act · GDPR Art. 22 | Policy enforcement layers + audit logging + human oversight gates |
| 02 | Data Poisoning | CRITICAL | OWASP LLM03 | Data provenance tracking + anomaly detection + sandbox model updates |
| 03 | Excessive Autonomy Risks | HIGH | OWASP Agentic Top 10 | HITL controls + agent registry + minimal permissions + kill-switch |
| 04 | Supply Chain Vulnerabilities | HIGH | OWASP LLM05 · Verizon DBIR 2025 | Vendor audits + artifact signing + AIBOM + OAuth token monitoring |
| 05 | Model Drift | MEDIUM | NIST AI RMF MEASURE | Continuous monitoring + drift detection + retraining governance |
| 06 | Model Inversion | HIGH | OWASP LLM06 · GDPR Art. 17 | Differential privacy + output filtering + adversarial probing + rate limiting |
| 07 | API Key & Credential Theft | CRITICAL | IBM X-Force 2026 | Secret rotation + vault storage + SPIFFE workload identity + usage anomaly detection |
| 08 | Unauthorized Tool Invocation | HIGH | OWASP LLM07 | Minimal tool surface + input validation + HITL for destructive calls + dry-run mode |
| 09 | Context Leakage | HIGH | EchoLeak CVE-2025-32711 | Context segmentation + DLP layers + treat retrieved content as untrusted + prompt logging |
Continuous Reassessment Is Not Optional
The nine threat vectors documented here were not theoretical when this model was written. Every documented incident cited — the Internet Archive breach, the UNC6395 OAuth attack, EchoLeak, OpenClaw, the 300,000 stolen ChatGPT credentials, the first AI-orchestrated cyber-espionage campaign — occurred in 2024–2026 in production enterprise environments. These are the operational conditions that every AI deployment team is working in right now.
Traditional security frameworks were not designed for AI’s unique threat surface. Data poisoning does not require hacking a server. Model inversion only requires API access. Context leakage can happen without any code execution. Excessive autonomy risk is created the moment an agent is granted broad permissions — before any attacker is involved. The controls that address these vectors — SPIFFE workload identity, policy-as-code enforcement, continuous drift monitoring, differential privacy, data provenance tracking — are AI-native security disciplines, not extensions of traditional perimeter defences.
Organisations that adopt a static, point-in-time security audit against this threat model will find their controls outdated before the next budget cycle. New attack vectors emerge with every new AI capability release — OWASP went from no LLM framework to a Top 10 for both LLMs and Agentic Applications in under 24 months. The only sustainable posture is continuous reassessment: quarterly at minimum, and immediately after any trigger event — a new AI vendor, a regulatory update, a new model deployment, or a threat intelligence report describing a new attack class.