Skip to content

Your First 90 Days as an AI Security Engineer

Series: AI Security in Practice
Pillar: 1: Foundations
Difficulty: Beginner
Author: Paul Lawlor
Date: 21 February 2026
Reading time: 18 minutes

A week-by-week playbook for the person who just landed the AI security role or was told “you own AI security now,” covering asset inventory, trust boundary mapping, red team testing, and presenting a risk register to leadership.


  1. Why Sequencing Matters More Than Knowledge
  2. Week 1: Inventory Every AI System and Shadow AI Usage
  3. Weeks 2–4: Map Trust Boundaries and Data Flows for the Highest-Risk Systems
  4. Month 2: Run Your First Red Team Exercise
  5. Month 3: Present a Risk Register to Leadership
  6. Ongoing: Benchmark Against the Maturity Model
  7. The 90-Day Checklist

1. Why Sequencing Matters More Than Knowledge

Section titled “1. Why Sequencing Matters More Than Knowledge”

You have been told “you own AI security now.” Perhaps it is a formal title change. Perhaps it is a sentence in a one-to-one with your CISO. Perhaps it is an email that says “we need someone to cover this” and nobody else raised their hand. However it happened, you are here, and the clock is ticking.

The temptation is to start learning everything. Read every paper on adversarial machine learning. Complete every course on LLM security. Build a mental model of every framework before touching anything in production. This is a mistake. Not because the knowledge is unimportant, but because it creates a false sense of progress while real risks go unaddressed.

A 2024 McKinsey survey found that 72% of organisations had adopted AI in at least one business function, up from roughly 50% in the years prior. 1 In most of those organisations, AI adoption outpaced the security team’s ability to assess, instrument, and govern it. By the time someone is assigned to own AI security, there is already a backlog of ungoverned systems, unreviewed integrations, and unknown shadow AI usage. The gap is not knowledge. The gap is action, sequenced correctly.

This playbook provides a week-by-week action plan for your first 90 days. It is not a job description. It is not a maturity model (that is Article 1.06 on this site). 2 It is the minimum viable set of actions, in the right order, to move your organisation from “nobody owns AI security” to “we have visibility, we have tested our defences, and we can articulate our risk posture to leadership.”

The sequencing follows a deliberate logic. You cannot map trust boundaries for systems you have not inventoried. You cannot red team systems whose data flows you have not mapped. You cannot present a credible risk register without evidence from both mapping and testing. Each phase builds on the one before it.

If you are an experienced security engineer or pentester pivoting into AI security, much of what you already know transfers directly: threat modelling, least privilege, input validation, and secure development lifecycle thinking all apply. 3 What is genuinely new is covered in the sections that follow and in the deeper articles they link to. If you are newer to security, the skills map in Article 1.09 on this site provides an honest triage of what to learn first and what can wait. [4]

The goal at the end of 90 days is not perfection. It is a defensible position: you know what AI systems exist, you know where the highest risks are, you have tested at least one system, and you can stand in front of leadership with a risk register and a plan. Everything else is iteration.


2. Week 1: Inventory Every AI System and Shadow AI Usage

Section titled “2. Week 1: Inventory Every AI System and Shadow AI Usage”

Your first week has one objective: build a complete picture of every AI system in use across the organisation. You cannot secure what you do not know about, and in most organisations, the official list of AI tools covers less than half of actual usage.

Start with what is known. Work with IT, procurement, and engineering leadership to build a list of every sanctioned AI system. This includes:

  • AI coding assistants (GitHub Copilot, Cursor, Amazon Q Developer, Windsurf, Devin)
  • Customer-facing LLM integrations (chatbots, search, recommendation engines, content generation)
  • Internal LLM tools (document summarisation, code review, data analysis)
  • Third-party AI APIs (OpenAI, Anthropic, Google Vertex AI, AWS Bedrock, Azure OpenAI)
  • Self-hosted models (Ollama, vLLM, TGI, Triton Inference Server)
  • AI features embedded in existing products (Salesforce Einstein, Microsoft 365 Copilot, ServiceNow AI)

For each system, record: the business owner, the data it processes, where the model runs (cloud API, self-hosted, on-device), what external services it calls, and whether a security review has been conducted. A spreadsheet is sufficient at this stage. Do not let the perfect be the enemy of the done.

The harder task is finding what nobody told you about. Shadow AI is the AI equivalent of shadow IT: tools adopted by individuals or teams without central approval. A 2024 Netskope report found that the average enterprise had employees using over a dozen distinct generative AI applications, many without IT awareness. 4

Three methods to discover shadow AI:

  1. Network traffic analysis. Work with your network team to review DNS logs and web proxy logs for traffic to known AI service domains: api.openai.com, api.anthropic.com, generativelanguage.googleapis.com, bedrock-runtime.*.amazonaws.com, *.openai.azure.com, and model hosting endpoints like huggingface.co. This gives you volume and frequency, not content.

  2. Endpoint telemetry. If you run an EDR or endpoint management solution, query for AI-related browser extensions, desktop applications (ChatGPT desktop, Claude desktop, Ollama), and IDE plugins.

  3. Survey the organisation. Send a short, non-punitive survey asking teams which AI tools they use, what data they provide to those tools, and what they use the output for. Frame it as “help us support you safely” rather than “tell us what you have been doing wrong.” You will get better response rates and more honest answers.

By the end of Week 1, you should have a single document listing every known and discovered AI system. For each, note: system name, type, data classification of inputs, hosting model, business owner, and whether a security review exists. This inventory is the foundation for everything that follows.

If the inventory reveals more than you expected (it usually does), resist the urge to immediately restrict everything. Banning tools without providing alternatives drives usage further underground. 5 Note the risks, prioritise, and address them in the phases that follow.


3. Weeks 2–4: Map Trust Boundaries and Data Flows for the Highest-Risk Systems

Section titled “3. Weeks 2–4: Map Trust Boundaries and Data Flows for the Highest-Risk Systems”

With the inventory complete, you now know what exists. The next three weeks are about understanding how the highest-risk systems work and where the security boundaries are. You will not have time to map every system in detail. Prioritise ruthlessly.

Rank the systems from your inventory using three criteria:

  1. Data sensitivity. Systems that process customer PII, financial data, source code, or classified information rank highest. A customer-facing chatbot that accesses a knowledge base of internal documents is higher risk than an internal tool that summarises public meeting notes.

  2. Agency and autonomy. Systems that can take actions (call APIs, execute code, send emails, modify databases) are higher risk than those that only generate text. An AI agent with tool-calling capability is a fundamentally different risk profile from a standalone chatbot. [7]

  3. Exposure. Internet-facing systems are higher risk than internal-only tools. Systems accessible to unauthenticated users are higher risk than those behind SSO.

Pick the top three to five systems. These are your focus for the next three weeks.

For each selected system, create a data flow diagram that answers:

  • What data enters the system? User prompts, uploaded documents, RAG-retrieved content, API responses, tool outputs. Identify which inputs come from trusted sources and which come from untrusted sources. In most LLM systems, RAG-retrieved content and tool outputs should be treated as untrusted. 6

  • What does the model have access to? System prompts, function definitions, database connections, API credentials, file system access. Map the model’s effective permissions. The principle of least privilege applies here exactly as it does in traditional systems, but many LLM deployments grant models far more access than they need. 7

  • What leaves the system? Model responses to users, function calls to external services, log entries, telemetry. Identify every egress path where data could be exfiltrated, either through direct model output or through side-channel tool invocations.

  • Where are the trust boundaries? The boundary between user input and system instructions is the most critical one in any LLM system, and it is the one that prompt injection exploits. Unlike traditional applications, LLMs have no architectural separation between instructions and data: everything is processed as an undifferentiated token stream. 8 Understanding this is essential for assessing where defences need to be strongest.

For each system, walk through the OWASP Top 10 for LLM Applications (2025) 7 and assess which risks apply. Not every risk applies to every system. A system with no tool-calling capability is not vulnerable to LLM01’s function-calling exploitation scenarios, but it may still be vulnerable to prompt injection that causes it to leak its system prompt or RAG content. A system with RAG is vulnerable to indirect prompt injection via poisoned documents. 9

Record each applicable risk, its current mitigation status (mitigated, partially mitigated, unmitigated), and the potential business impact. This mapping feeds directly into the risk register you will build in Month 3.

At the end of this phase, you will have a trust boundary diagram and OWASP risk mapping for each of your highest-risk systems. You will also have a list of gaps: missing input validation, excessive model permissions, absent logging, unmonitored tool calls. Do not attempt to fix everything immediately. Document the gaps, assign preliminary severity ratings, and carry them forward. The risk register in Month 3 is where they become actionable priorities.

The MITRE ATLAS framework 10 provides additional structure for this mapping. If your organisation already uses MITRE ATT&CK for traditional threat modelling, ATLAS provides a complementary taxonomy for AI-specific attack techniques that maps naturally to the same workflow.


4. Month 2: Run Your First Red Team Exercise

Section titled “4. Month 2: Run Your First Red Team Exercise”

You have the inventory. You have the trust boundary maps. Now it is time to test whether the defences actually hold. Month 2 is dedicated to your first AI red team exercise.

It may feel premature to attack your own systems when you have not finished building defences. It is not. Red teaming at this stage serves three purposes. First, it converts the theoretical risks from your OWASP mapping into concrete, demonstrable findings that leadership and engineering teams will take seriously. A documented prompt injection that extracts a customer’s credit card number from a RAG pipeline is worth more than a dozen risk matrix slides. Second, it builds your own hands-on intuition for how AI attacks work, which no amount of reading can substitute for. Third, it establishes the practice of adversarial testing before production, which is the habit that matters most in the long run.

Select one system from your trust boundary maps. The ideal first target is a customer-facing LLM integration (a chatbot, a search assistant, a document summarisation tool) that processes real data. Avoid starting with internal-only tools: the findings will be less compelling to leadership, and the incentive to fix them will be weaker.

If you do not have a suitable production system to test (or if testing production carries too much risk), set up a local lab environment. Article 1.03 on this site walks through building a home AI security lab with Ollama, PyRIT, and Garak. 11 You can stand up a vulnerable chatbot backed by a RAG pipeline in an afternoon, and the attacks you practise locally will transfer directly to production testing.

Two open-source tools cover the ground you need for a first exercise:

PyRIT (Python Risk Identification Toolkit), developed by Microsoft’s AI Red Team, provides orchestrated attack strategies against LLM endpoints. It supports multi-turn attacks, automated scoring of responses, and integration with multiple model providers. 12 Article 2.01 on this site provides a complete setup and attack guide. 13 Start with PyRIT’s prompt injection orchestrators: they will systematically test whether your system’s guardrails can be bypassed to extract system prompts, access restricted data, or invoke unauthorised functions.

Garak, developed by NVIDIA, is an LLM vulnerability scanner that runs a library of probes against a target endpoint. 14 It covers prompt injection, data leakage, toxicity, hallucination, and more. Garak is closer to a traditional vulnerability scanner: you point it at a target and it runs a battery of tests. See PyRIT (Article 2.01) for red teaming tool setup; Garak is covered in the same tooling ecosystem.

For a first exercise, run Garak’s standard probe set against your target to establish a baseline, then use PyRIT for targeted, multi-turn attacks against the vulnerabilities Garak identifies.

Structure the exercise as follows:

  1. Scope and rules of engagement. Define what is in scope (the target system, its data, its connected services) and what is out of scope (other production systems, data destruction, denial of service). Get written approval from the system owner and, if applicable, the CISO. This is standard practice for any penetration test and applies equally to AI red teaming.

  2. Baseline scan. Run Garak against the target endpoint. Review the results for critical findings: prompt injection successes, system prompt extraction, data leakage, and any unauthorised function calls. Record pass/fail rates for each probe category.

  3. Targeted attacks. Use PyRIT to attempt multi-turn prompt injection scenarios based on your trust boundary analysis. If the system has RAG, test for indirect prompt injection by planting adversarial content in documents the system retrieves. 9 If the system has tool-calling capability, test whether you can induce it to call tools with attacker-controlled parameters. [7]

  4. Document findings. For each finding, record: the attack technique used, the OWASP LLM Top 10 risk it maps to, the MITRE ATLAS technique ID, the steps to reproduce, the business impact, and a recommended remediation. Use the same format your organisation uses for traditional penetration test findings. Consistency with existing processes makes adoption easier.

The findings from this exercise serve two purposes. Immediate remediation: work with the engineering team that owns the system to address critical findings. Longer-term evidence: the findings feed directly into the risk register you will present to leadership in Month 3. A risk register backed by concrete, reproducible findings is categorically more persuasive than one filled with theoretical risks.


5. Month 3: Present a Risk Register to Leadership

Section titled “5. Month 3: Present a Risk Register to Leadership”

The final phase of your first 90 days culminates in the artefact that justifies your role and secures ongoing investment: an AI risk register, presented to leadership with findings, context, and a prioritised remediation plan.

An AI risk register follows the same structure as any enterprise risk register, extended with AI-specific categories. The NIST AI RMF Implementation Guide provides risk register templates and worked examples. 15 For your first iteration, each entry should contain:

FieldDescription
Risk IDUnique identifier (e.g., AI-RISK-001)
Risk categoryMap to OWASP LLM Top 10 (e.g., LLM01 Prompt Injection, LLM06 Excessive Agency) 7
Affected systemThe specific system from your inventory
DescriptionWhat could happen, in business terms
LikelihoodBased on your red team findings and exposure assessment
ImpactFinancial, reputational, regulatory, operational
Current controlsWhat is already in place
Residual riskRisk level after current controls
Recommended actionSpecific remediation with estimated effort
OwnerThe person accountable for remediation

Populate the register from three sources: the OWASP risk mapping from Weeks 2-4, the concrete findings from your Month 2 red team exercise, and any gaps identified during the trust boundary analysis. Aim for 10 to 20 entries for a first iteration. Fewer than 10 suggests you have not looked hard enough. More than 30 risks diluting the message for a first presentation.

The NIST AI Risk Management Framework 16 provides the governance structure that gives your risk register legitimacy. Its four functions (Govern, Map, Measure, Manage) map directly to what you have done so far:

  • Govern: You have established ownership of AI security (that is you).
  • Map: You have inventoried AI systems and mapped trust boundaries.
  • Measure: You have run a red team exercise and quantified findings.
  • Manage: The risk register and remediation plan are your management output.

Framing your work in NIST AI RMF terms is particularly useful if your organisation already uses NIST frameworks for other domains. It demonstrates that AI security is not a separate discipline requiring entirely new processes, but an extension of existing risk management practices into a new domain.

The leadership presentation is not a technical briefing. It is a business case. Structure it as follows:

  1. Scope of AI usage. Present the inventory numbers. How many AI systems are in use? How many were unknown before your audit? What data classifications do they process? These numbers establish the scale of the problem.

  2. Key findings from red team testing. Select three to five findings that demonstrate real, concrete risk. A successful prompt injection that extracted customer data. An AI agent with excessive permissions to production databases. A RAG pipeline with no input validation on retrieved documents. Choose findings that connect to business outcomes, not technical curiosities.

  3. The risk register. Present the top five risks by residual risk score. For each, state the risk in business terms, the current control status, and the recommended remediation.

  4. The ask. Be specific about what you need. Budget for tooling (guardrails, monitoring, red team tools). Engineering time for remediation. Policy decisions (acceptable use policy, AI procurement requirements). Headcount if the workload justifies it. Quantify where possible: “Deploying guardrails on our three customer-facing LLM endpoints requires 40 engineering hours and a monthly service cost of approximately £2,000.”

  5. The roadmap. Show where you are against the AI Security Maturity Model (Article 1.06 on this site). 2 Most organisations at this stage will be at Level 1 (Ad Hoc) or early Level 2 (Developing). Present a 12-month plan to reach Level 3 (Defined), with quarterly milestones.

Do not present AI security as a reason to block AI adoption. You will lose the argument and your credibility. Frame every risk with a remediation, and frame every remediation as enabling safer adoption. The organisation is going to use AI regardless of what you say. Your job is to make that usage defensible.

Do not overstate findings. If a prompt injection only works under contrived conditions, say so. Credibility is your most important asset in the first 90 days, and overstating a risk will cost you trust that takes months to rebuild.


6. Ongoing: Benchmark Against the Maturity Model

Section titled “6. Ongoing: Benchmark Against the Maturity Model”

The 90-day plan gets you to a defensible starting position. What follows is iteration, measured against the AI Security Maturity Model described in Article 1.06 on this site. 2

If you have followed this playbook, your organisation has moved from Level 1 (Ad Hoc) to early Level 2 (Developing) on the maturity model. Specifically:

  • Governance: You have a named AI security owner (you), an AI system inventory, and the beginning of a risk register. You do not yet have a formal AI security policy or framework alignment.
  • Technical Controls: You have identified gaps in input validation, access controls, and guardrails, but most are not yet deployed. You have tools (PyRIT, Garak) for testing but they are not yet integrated into CI/CD pipelines.
  • Monitoring and Detection: You have visibility into which AI systems exist, but structured logging of LLM interactions is likely inconsistent or absent.
  • Incident Response: You may have added basic AI scenarios to existing playbooks, but a dedicated AI incident response plan is not yet in place. See The NIST AI RMF Implementation Guide for risk management and incident response structure. 17
  • Workforce and Culture: You have begun building your own expertise. The rest of the security team and the broader organisation are at the awareness stage.

With the foundation laid, the next quarter should focus on:

  1. Publish an AI acceptable use policy. Define which tools are approved, what data may be shared, and who to contact with questions. This single document is the most impactful governance action you can take. 16

  2. Deploy guardrails on your highest-risk LLM endpoints. Start with the systems identified in your trust boundary mapping. Options include cloud-native services like AWS Bedrock Guardrails, open-source tools like NeMo Guardrails, or commercial products like Lakera Guard. Article 3.01 on this site compares the major options. 18

  3. Implement structured logging. Every LLM interaction in production should log: the prompt, the response, any tool calls, token counts, and latency. Observability platforms like LangFuse provide purpose-built infrastructure for this. 19

  4. Integrate red team scanning into deployment pipelines. Move from manual, one-off testing to automated scanning. Run Garak probes as a gate in your CI/CD pipeline for any system that serves an LLM endpoint. 14

  5. Deliver targeted AI security training. Developers using AI coding assistants need specific guidance on reviewing AI-generated code for vulnerabilities. The security team needs hands-on experience with AI attack techniques. Article 1.09 on this site provides a skills map to structure this training. [4]

Repeat the maturity self-assessment from Article 1.06 every quarter. 2 Track the per-dimension scores over time. The dimensions with the lowest scores tell you where to invest next. Share the results with leadership alongside your risk register update. Progress measured against a framework is more compelling than a list of activities completed.

The threat landscape will evolve between assessments. New attack techniques will emerge. MITRE ATLAS adds new techniques regularly (14 new agentic AI techniques were added in October 2025 alone). 10 The OWASP LLM Top 10 will update. Your maturity assessment must account for new risks, not only progress against old ones.


This checklist consolidates every deliverable from the playbook. Use it as a tracking tool throughout the 90 days and as evidence of progress when reporting to leadership.

  • List all sanctioned AI tools and systems with business owners
  • Analyse network traffic for shadow AI service endpoints
  • Review endpoint telemetry for AI-related applications and extensions
  • Survey teams for undisclosed AI tool usage
  • Compile consolidated AI system inventory with data classifications
  • Identify systems processing sensitive data (PII, financial, source code, classified)

Weeks 2–4: Trust boundaries and risk mapping

Section titled “Weeks 2–4: Trust boundaries and risk mapping”
  • Select top 3–5 highest-risk systems based on data sensitivity, agency, and exposure
  • Create data flow diagrams for each selected system
  • Map model inputs, permissions, and output/egress paths
  • Identify trust boundaries (user input vs. system instructions, trusted vs. untrusted data)
  • Walk each system through the OWASP LLM Top 10 risk categories
  • Map applicable MITRE ATLAS techniques to each system
  • Document gaps in controls with preliminary severity ratings
  • Define scope and rules of engagement for the target system
  • Obtain written approval from system owner
  • Install and configure PyRIT and Garak (or set up local lab per Article 1.03)
  • Run Garak baseline scan against target endpoint
  • Execute targeted multi-turn attacks with PyRIT
  • Test for indirect prompt injection via RAG pipeline (if applicable)
  • Test for unauthorised tool invocation (if applicable)
  • Document each finding with reproduction steps, OWASP/ATLAS mapping, and business impact
  • Share critical findings with engineering team for immediate remediation

Month 3: Risk register and leadership presentation

Section titled “Month 3: Risk register and leadership presentation”
  • Build AI risk register with 10–20 entries from all prior phases
  • Assign likelihood and impact ratings based on evidence
  • Map risk register to NIST AI RMF functions (Govern, Map, Measure, Manage)
  • Prepare leadership presentation: scope, findings, top 5 risks, the ask, the roadmap
  • Complete AI Security Maturity Model self-assessment (Article 1.06)
  • Present to CISO / leadership and secure agreement on next-quarter priorities
  • Publish initial AI acceptable use policy (or confirm existing policy is adequate)

The following articles on this site provide deeper guidance for each phase of the playbook:

PhaseArticleWhy it helps
Getting started1.03 Building a Home AI Security LabHands-on lab environment for learning and testing
Getting started1.08 AI Security vs Application SecurityWhat transfers from your existing security skills
Getting started1.09 The AI Security Skills MapWhat to learn first and what can wait
Inventory6.06 The Shadow AI ProblemDetecting unauthorised AI tool usage at scale
Trust boundaries1.02 The AI Threat Landscape: OWASP LLM Top 10Practitioner walkthrough of each OWASP risk
Trust boundaries2.06 The MITRE ATLAS PlaybookMapping AI attacks to the ATT&CK framework
Red teaming2.01 PyRIT from Zero to Red TeamComplete PyRIT setup and attack guide
Red teaming2.09 Garak: Open-Source LLM Vulnerability ScanningGarak setup and probe selection
Risk register5.05 Building an AI Risk RegisterTemplate and worked examples
Incident response5.06 AI Incident ResponseThe playbook your SOC needs
Maturity model1.06 The AI Security Maturity ModelSelf-assessment and level-by-level roadmap
Programme build1.10 Building an AI Security Programme from ZeroThe full programme construction manual

  1. McKinsey & Company, “The state of AI in early 2024: Gen AI adoption spikes and starts to generate value” (May 2024). Available at: https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai-2024

  2. AI Security in Practice, “The AI Security Maturity Model”, Article 1.06 on this site. 2 3 4

  3. AI Security in Practice, “AI Security vs Application Security”, Article 1.08 on this site.

  4. Netskope, “Cloud and Threat Report: AI Apps in the Enterprise” (2024). Available at: https://www.netskope.com/netskope-threat-labs/cloud-threat-report

  5. AI Security in Practice, “The Shadow AI Problem”, Article 6.06. See AI Supply Chain Attacks for related content on unauthorised AI usage risks.

  6. AI Security in Practice, “The RAG Trap”, Article 3.11 on this site.

  7. OWASP, “Top 10 for LLM Applications 2025”. Available at: https://genai.owasp.org/ 2 3

  8. AI Security in Practice, “How LLMs Work: A Security Engineer’s Guide to Tokenisation, Attention, and RLHF”, Article 1.01 on this site.

  9. Greshake, K. et al., “Not what you’ve signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection” (2023). Available at: https://arxiv.org/abs/2302.12173 2

  10. MITRE, “ATLAS: Adversarial Threat Landscape for AI Systems”. Available at: https://atlas.mitre.org/ 2

  11. AI Security in Practice, “Building a Home AI Security Lab”, Article 1.03 on this site.

  12. Microsoft, “PyRIT: Python Risk Identification Toolkit for generative AI”. Available at: https://github.com/Azure/PyRIT

  13. AI Security in Practice, “PyRIT from Zero to Red Team”, Article 2.01 on this site.

  14. NVIDIA, “Garak: LLM Vulnerability Scanner”. Available at: https://github.com/NVIDIA/garak 2

  15. AI Security in Practice, “The NIST AI RMF Implementation Guide” provides risk register templates and worked examples.

  16. NIST, “AI Risk Management Framework (AI RMF 1.0)”, NIST AI 100-1, January 2023. Available at: https://www.nist.gov/itl/ai-risk-management-framework 2

  17. AI Security in Practice, “AI Incident Response: The Playbook Your SOC Doesn’t Have Yet”, Article 5.06 on this site.

  18. AI Security in Practice, “Guardrails Engineering”, Article 3.01 on this site.

  19. LangFuse, “Open Source LLM Observability”. Available at: https://langfuse.com/