
What should worry you???
How about the AI systems that are already in production, making decisions that affect customers and revenue? How about threat models that still assume software behaves predictably?
Deterministic thinking will get you in trouble with systems that learn from data, adapt to new inputs, and generate outputs that are not strictly bounded by static logic.
You are rolling out copilots, recommendation engines, fraud detection models, and automated workflows at speed, yet the review process behind them often looks the same as it did five years ago. Traditional threat modeling was built around fixed trust boundaries, known data flows, and clearly defined misuse cases. AI systems is different from that foundation. Model manipulation, prompt injection, data poisoning, and output-driven abuse expand the attack surface in ways older frameworks were never designed to capture.
Traditional threat modeling works because classic software behaves like a closed system. You map components, draw trust boundaries, list entry points, and reason about what an attacker can do when they control an input or a dependency. That logic holds when your software is deterministic, inputs are bounded, and the execution path is mostly predictable. AI systems punch holes in all three assumptions, and that changes what coverage even means in a threat model.
In conventional software, the same input reliably produces the same output unless state changes. You can test edge cases, enforce invariants, and treat unexpected behavior as a defect in code or configuration. AI systems behave differently because the model is a statistical machine, not a rules engine. Even when you send the same prompt, you can get variable outputs based on sampling, context length, tool selection behavior, retrieval results, model updates, and hidden system prompts. That variability is not a bug. It is the design.
What changes in practice is what you must treat as part of the security boundary:
Most organizations lean on STRIDE-style thinking or architecture workshops that center on familiar categories like injection, privilege escalation, data exposure, and abuse of service. Those still matter for AI systems because you still ship APIs, run services, handle auth, and store data. The problem is that AI introduces additional threat categories that sit above the code layer and below the user layer, living in the interaction between model behavior, data, and orchestration.
Here are the categories that show up fast once AI is in the loop:
Traditional models often miss these because they focus on the app boundary, while AI risks often sit in the content boundary, the orchestration boundary, and the decision boundary.
A threat model that reviews only application code and APIs is not modeling AI risk, because the attack surface includes the model’s context, the retrieval layer, the tools the model can call, and the downstream systems that trust its outputs. You need to treat prompts, retrieved content, and orchestration logic as security-critical components, then build controls that prevent instruction hijacking, limit data exposure through inference, constrain tool access, and detect behavior drift in production.
Most threat modeling sessions for AI systems still look like a standard application review. The team walks through services, APIs, auth, data stores, then calls it done. That approach misses the point because AI introduces structural layers that behave like separate subsystems, each with its own trust boundaries, failure modes, and attacker incentives.
Here’s a clean way to map the layers so you can reason about risk without guessing.
Traditinal reviews miss risks because they stay anchored to code paths and endpoint threats, while AI failures often come from how these layers interact. Once you map layers separately, the gaps become obvious and they look familiar in hindsight.
What typically gets missed:
This is exactly why AI-specific frameworks exist. OWASP LLM Top 10 calls out risks that show up in real GenAI deployments, like prompt injection and insecure output handling. MITRE ATLAS focuses on adversarial tactics and techniques against AI systems, which helps teams think beyond web-app exploit patterns. NIST AI RMF forces governance, measurement, and management of AI risk across the lifecycle, which pushes threat modeling toward repeatable, auditable control coverage instead of one-time workshops.
Traditional threat modeling is an event. A workshop happens, a diagram gets updated, a PDF gets stamped, and everyone moves on. AI systems do not behave like that, because they keep changing in ways that affect risk even when no one touches application code, and those changes rarely trigger the kind of review your program is built around.
Once AI is part of your product or internal operations, the system you are securing includes moving parts that update on their own cadence, owned by different teams, and deployed through different pipelines.
The most common change vectors look like this:
Threat models are built on assumptions. You assume which data sources are trusted, which tools exist, what the model is allowed to do, and where outputs flow. In AI systems, those assumptions age fast because the system boundary is not stable.
Security drift shows up in a few predictable ways:
At that point, you are not managing risk, you are managing a document.
Regulatory pressure is moving toward continuous AI risk management, not one-time assessments. Boards want defensible oversight, which means you need to show how risk is monitored as the system changes, how controls are enforced, and how exceptions are managed over time. A static PDF threat model can describe what the system looked like at review time, but it cannot prove that your assumptions still hold after weekly index updates, prompt revisions, new connectors, or model refreshes.
Traditional threat modeling broke because it was built for deterministic, code-driven systems, and AI systems learn from data, adapt through retraining and prompt changes, generate probabilistic outputs, and directly influence business decisions.
Security has to move from code-centric reviews to system-behavior-centric analysis. That means treating the model, data pipelines, orchestration logic, and runtime integrations as separate trust boundaries, then continuously assessing how changes in prompts, embeddings, connectors, and retraining cycles shift your exposure.
Start by auditing your current AI threat modeling approach and identify which layers are not being modeled and whether your reviews are continuous or static.
At we45, we help security leaders do exactly that through deep AI architecture reviews, adversarial testing, and continuous threat modeling programs designed for systems that evolve every week.
If your threat model stops at the application layer, that is exactly where your AI exposure begins.
Traditional threat modeling assumes software is a deterministic, closed system with fixed trust boundaries and predictable logic. AI systems are fundamentally different because they are probabilistic, learn from data, and adapt to new inputs. This breaks the core assumptions of older frameworks, which were not designed to capture risks like prompt injection or data poisoning.
In conventional software, the same input yields the same output. AI models, being statistical machines, can produce variable outputs even from the same prompt due to factors like sampling, context length, model updates, and hidden system prompts. This variability means security must be enforced not just through code, but through continuous monitoring, guardrails, and policy enforcement in production.
AI systems introduce new, behavioral threat categories that traditional models often miss. These include prompt injection and instruction hijacking, context poisoning, model inversion and sensitive data inference, data extraction via tool use, and output manipulation leading to downstream abuse. These risks live in the content, orchestration, and decision boundaries, not just the application boundary.
The AI attack surface is layered and includes four main subsystems, each with its own risks: Model layer: Base model selection, fine-tuning artifacts, weights, and serving configuration. Data layer: Training data, embeddings generation, RAG knowledge bases, and feedback loops. Orchestration layer: System prompts, tool integrations, API connectors, and guardrails logic. Runtime layer: Inference APIs and gateways, external content ingestion, and live decision logic.
AI systems are subject to constant change, known as security drift, even without application code updates. Changes can occur through model retraining, updates to RAG indexes, tweaks to system prompts, adding new external data sources, and expanding tool integrations. A static threat model quickly becomes obsolete because its underlying assumptions no longer match the evolving system boundary and capabilities.
AI threat modeling requires a structural shift from code-centric reviews to continuous, system-behavior-centric analysis. This involves treating the model, data pipelines, orchestration logic, and runtime integrations as separate trust boundaries and continuously assessing how changes in these components shift the system’s exposure and risk posture.