Back to blog
    Shadow AI Audit Checklist: Find Every Unauthorized AI Tool in Your Organization
    shadow-aiauditenterprise-securitychecklistcompliancesegment:enterprise

    Shadow AI Audit Checklist: Find Every Unauthorized AI Tool in Your Organization

    A step-by-step audit process to discover unauthorized AI tools in your organization. Covers network traffic analysis, browser extension audits, SaaS spend analysis, employee surveys, DLP reviews, and API key audits — with a 25-item checklist you can use immediately.

    EErtas Team·

    You cannot secure what you cannot see. Before writing policies, deploying monitoring tools, or rolling out sanctioned AI alternatives, you need an accurate picture of what AI tools are currently being used in your organization, by whom, and with what data.

    This is the shadow AI audit. It is a structured, repeatable process that produces a complete inventory of unauthorized AI usage — tools, users, data categories, and risk levels. This article provides the full process, including a 25-item checklist you can adapt to your organization immediately.

    Organizations with 200–1,000 employees interact with an average of 45 distinct AI websites monthly. In larger organizations, that number climbs to 72. The audit will almost certainly reveal more tools than you expect.


    Before You Start: Set the Scope

    Define three things before beginning:

    Time window. Audit the last 90 days. This captures habitual usage patterns without going so far back that the data is stale. AI tool adoption is accelerating — usage patterns from six months ago may not reflect current behavior.

    Organization scope. Decide whether this is a company-wide audit or a department-first approach. If you are resource-constrained, start with departments that handle the most sensitive data: legal, HR, finance, engineering, and customer-facing teams.

    Desired outcome. The audit should produce three deliverables: (1) an inventory of all AI tools in use, (2) a risk classification of each tool based on data sensitivity, and (3) a list of immediate remediation actions ranked by risk.


    Phase 1: Network Traffic Analysis

    Network logs are your most objective data source. They show what actually happened, not what employees remember or choose to disclose.

    What to look for

    Monitor outbound connections to known AI service domains. The following list covers the major services as of March 2026, but new tools appear monthly — update this list quarterly.

    AI ServiceDomains to Monitor
    OpenAI / ChatGPTchat.openai.com, api.openai.com, cdn.openai.com
    Anthropic / Claudeclaude.ai, api.anthropic.com, console.anthropic.com
    Google Geminigemini.google.com, generativelanguage.googleapis.com, aistudio.google.com
    Perplexityperplexity.ai, api.perplexity.ai
    xAI / Grokgrok.x.ai, x.com/i/grok
    Mistralchat.mistral.ai, api.mistral.ai
    Coherecoral.cohere.com, api.cohere.ai
    Hugging Facehuggingface.co, api-inference.huggingface.co
    Poepoe.com
    Character.AIcharacter.ai
    Jasperapp.jasper.ai
    Copy.aiapp.copy.ai
    Notion AInotion.so (look for AI-specific API calls)
    GitHub Copilotcopilot.github.com, copilot-proxy.githubusercontent.com
    Cursor AIcursor.sh, api2.cursor.sh
    Codeiumcodeium.com, server.codeium.com

    How to extract this data

    • Firewall logs: Export connection logs from your firewall or UTM appliance. Filter for the domains above. Most enterprise firewalls (Palo Alto, Fortinet, Sophos) can export to CSV or SIEM.
    • DNS query logs: If you run internal DNS, query logs show every domain lookup regardless of whether the connection was completed. This catches tools that employees tried to access even if the firewall blocked them.
    • Proxy logs: If you route web traffic through a proxy (Zscaler, Netskope, etc.), you have full URL-level visibility including HTTPS inspected traffic.
    • SIEM aggregation: If you have a SIEM (Splunk, Sentinel, Elastic), create a dashboard that aggregates connections to the domains above, grouped by source IP → user mapping.

    What this tells you

    Network analysis gives you the tools in use, the frequency of use, and the users (via IP-to-user mapping from your directory service). It does not tell you what data was sent — that requires DLP inspection, which is covered in Phase 5.


    Phase 2: Browser Extension Audit

    AI-powered browser extensions are a significant and often overlooked vector. Employees install extensions that read page content, capture form data, or intercept clipboard content — and send it to external AI services for processing.

    What to audit

    • Chrome extension inventory: Use Google Admin Console (for managed Chrome) or endpoint management tools (Intune, Jamf) to pull a list of installed extensions across all managed devices.
    • Edge extension inventory: Same process via Microsoft Endpoint Manager.
    • Extension categories to flag: Any extension with permissions for "Read and change all your data on all websites," "Read your browsing history," or "Modify data you copy and paste."

    Common AI extensions to search for

    • ChatGPT browser extensions (multiple third-party variants)
    • Merlin (GPT-4 in browser)
    • Monica AI
    • Sider AI
    • MaxAI.me
    • UseChatGPT.AI
    • Compose AI
    • Writesonic
    • Grammarly (AI features — may be sanctioned, verify)
    • Otter.ai (meeting transcription — often installed individually)

    Risk assessment

    Extensions that have "read all site data" permissions can capture data from internal web applications (your CRM, ERP, intranet, code repositories) and send it to external servers. This is a higher risk than employees manually pasting data into ChatGPT, because it happens passively and continuously.


    Phase 3: SaaS Spend Analysis

    Individual AI subscriptions show up in expense reports and on corporate credit cards. This is a simple but effective detection method.

    Where to look

    • Expense reports: Search for reimbursement requests containing "OpenAI," "ChatGPT," "Claude," "Anthropic," "Jasper," "Copy.ai," "Midjourney," "Runway," or "AI" in the description.
    • Corporate credit card statements: Run the same keyword search across all corporate card transactions for the last 90 days.
    • Procurement system: Check if any team has submitted purchase requests for AI tools outside your standard procurement process.
    • App store charges: For organizations that reimburse mobile app purchases, check for AI app subscriptions (ChatGPT Plus at $20/month is the most common).

    What this tells you

    SaaS spend analysis identifies the employees who are invested enough in AI tools to pay for them. These are typically your heaviest users and the ones sending the most data. They are also, importantly, the ones who will have the strongest opinions about what a sanctioned alternative needs to do.


    Phase 4: Employee Survey

    Network logs tell you what tools are being used. An employee survey tells you why they are being used and what data is being processed. Both are necessary.

    Survey design principles

    • Anonymous. If employees fear punishment, they will not disclose usage. Make the survey explicitly anonymous and frame it as a discovery exercise, not an investigation.
    • Specific. Do not ask "Do you use AI tools?" Ask "In the last 30 days, have you used any of the following tools for work-related tasks?" and list specific tools.
    • Data-focused. Ask what categories of data they have used with AI tools: source code, customer data, financial data, HR data, legal documents, internal communications, meeting notes.

    Sample questions

    1. Which AI tools have you used for work-related tasks in the last 30 days? (Select all that apply — list specific tools)
    2. How frequently do you use AI tools for work? (Daily / Several times per week / Weekly / Monthly / Rarely)
    3. What types of work tasks do you use AI tools for? (Writing/editing, code development, data analysis, research, summarization, translation, other)
    4. What types of company data have you input into AI tools? (Source code, customer information, financial data, HR/personnel data, legal documents, meeting notes, internal communications, none, other)
    5. Do you use a personal account or a work-provided account for AI tools?
    6. What prevents you from using company-approved AI tools? (No approved tools exist, approved tools are too limited, approved tools are too slow, did not know approved tools existed, other)
    7. If the company provided an internal AI assistant with equivalent capabilities, would you use it instead of external tools?

    Question 6 is critical. It tells you why the sanctioned alternative (if you have one) is not being adopted, or confirms that you do not have one.


    Phase 5: DLP Policy Review

    Most organizations have Data Loss Prevention policies configured for email, USB, and cloud storage. Very few have extended those policies to cover AI tool usage.

    Audit your current DLP configuration

    • Does your DLP policy monitor browser uploads? Specifically, does it inspect content being posted via form submissions or API calls to AI service domains?
    • Does your DLP policy cover clipboard operations? Copy-paste into a browser-based AI tool is the most common data exfiltration method for shadow AI.
    • Are AI service domains in your DLP monitoring scope? Most DLP solutions can be configured to flag or block uploads to specific domains. Check if the AI domains from Phase 1 are included.
    • What data classification categories trigger alerts? Ensure that your DLP classifications (PII, PHI, financial data, source code, legal privileged) are applied to AI tool destinations, not just email and cloud storage.

    Gap assessment

    If your DLP does not cover AI tool uploads, you have a monitoring gap. Quantify it: based on Phase 1 network data, how many connections per day are being made to AI services, and what percentage of those could involve sensitive data based on the user roles making those connections?


    Phase 6: API Key Audit

    Developers and technical teams may be using AI services programmatically, embedding API keys in code repositories, environment files, CI/CD pipelines, or scripts. This usage is harder to detect through network monitoring because it may run on development machines, CI servers, or cloud infrastructure.

    • Code repositories: Search for strings like sk-, sk-proj-, OPENAI_API_KEY, ANTHROPIC_API_KEY, GOOGLE_API_KEY across all internal Git repositories (GitHub, GitLab, Bitbucket).
    • Environment files: Search for .env, .env.local, .env.production files containing AI service API keys.
    • CI/CD secrets: Review stored secrets in GitHub Actions, GitLab CI, Jenkins, or your CI/CD platform for AI service credentials.
    • Secrets scanning tools: If you use a secrets scanner (GitGuardian, TruffleHog, gitleaks), check if AI API key patterns are included in the scanning rules.
    • Cloud infrastructure: Check AWS Secrets Manager, Azure Key Vault, or GCP Secret Manager for AI service credentials that may have been provisioned outside standard procurement.

    Risk assessment

    API key usage typically involves automated or semi-automated data processing, which means higher volumes of data are being sent to external AI services. A developer running a script that processes 10,000 customer records through the OpenAI API is a higher-volume exposure than the same developer manually pasting a few records into ChatGPT.


    The Master Checklist

    Use this checklist to track audit completion. Each item should be marked complete with findings documented.

    #Audit ItemPhaseStatus
    1Export firewall/proxy logs for last 90 daysNetwork
    2Filter logs for all known AI service domainsNetwork
    3Map source IPs to user accounts via directory serviceNetwork
    4Count unique users and frequency per AI serviceNetwork
    5Check DNS query logs for AI domains not in firewall logsNetwork
    6Pull installed browser extension inventory from all managed devicesExtensions
    7Flag extensions with "read all site data" or clipboard permissionsExtensions
    8Cross-reference flagged extensions against known AI extension listExtensions
    9Search expense reports for AI tool reimbursements (last 90 days)SaaS Spend
    10Search corporate card transactions for AI service chargesSaaS Spend
    11Check procurement system for unapproved AI tool requestsSaaS Spend
    12Check mobile app reimbursements for AI subscriptionsSaaS Spend
    13Design and distribute anonymous employee AI usage surveySurvey
    14Analyze survey results: tools used, data types, frequencySurvey
    15Identify departments with highest shadow AI adoptionSurvey
    16Identify data categories most commonly pasted into AI toolsSurvey
    17Verify DLP policies cover browser uploads to AI domainsDLP
    18Verify DLP covers clipboard/paste operations to AI sitesDLP
    19Confirm AI service domains are in DLP monitoring scopeDLP
    20Confirm sensitive data classifications apply to AI egressDLP
    21Scan all code repositories for AI API keysAPI Keys
    22Search .env files across development environmentsAPI Keys
    23Audit CI/CD secrets for AI service credentialsAPI Keys
    24Check cloud secret stores for AI API credentialsAPI Keys
    25Compile findings into risk-ranked inventory with remediation prioritiesFinal

    Interpreting the Results

    After completing all six phases, you will have three categories of findings:

    Category 1: Known sanctioned tools with policy gaps

    These are tools the organization has approved (e.g., GitHub Copilot, Grammarly) but where the usage exceeds the approved scope. For example, Copilot is approved for code suggestions but developers are also pasting customer data into ChatGPT for debugging. The tool is sanctioned; the usage pattern is not.

    Action: Update policies to clarify approved vs. unapproved usage of sanctioned tools. Update DLP rules to enforce data category restrictions.

    Category 2: Unsanctioned tools with low data sensitivity

    These are consumer AI tools being used for tasks that do not involve sensitive data — rephrasing emails, generating meeting agenda templates, brainstorming. The tools are not approved, but the data risk is low.

    Action: Decide whether to sanction these tools (if the cost is minimal and the risk is low) or redirect this usage to an internal alternative. Do not over-react to low-risk usage — treating benign productivity usage as a security incident damages trust and drives the behavior underground.

    Category 3: Unsanctioned tools with high data sensitivity

    This is the critical finding. Sensitive data — source code, customer PII, financial data, legal documents — flowing into consumer AI tools through personal accounts with no organizational controls.

    Action: This requires immediate remediation. The priority order is:

    1. Deploy a sanctioned on-premise alternative for the specific use cases identified. If developers are pasting code into ChatGPT, they need an internal AI coding assistant. If legal teams are summarizing contracts, they need an internal AI document analysis tool. See How to Build a Sanctioned AI Alternative to ChatGPT for Your Enterprise.

    2. Update DLP policies to block or alert on sensitive data uploads to unsanctioned AI domains.

    3. Communicate with affected teams — not punitively, but to explain the risk and introduce the sanctioned alternative.

    4. Establish ongoing monitoring to detect new shadow AI tool adoption before it reaches the scale you just discovered.


    After the Audit: Building the Remediation Plan

    The audit is not the end — it is the input for a remediation plan. That plan should address four areas:

    Sanctioned alternatives. For every high-risk usage pattern identified, determine what sanctioned tool or internal deployment will replace it. The goal is not to eliminate AI usage — it is to move AI usage onto infrastructure you control.

    Policy. Draft or update your AI usage policy to reflect what the audit revealed. A policy that does not address the actual tools and use cases employees are using is a policy that will be ignored.

    Monitoring. Implement continuous monitoring for the six categories audited above. The audit is a point-in-time snapshot; monitoring provides ongoing visibility. Set a quarterly cadence for repeating the full audit.

    Training. Use the audit findings (anonymized) to build training materials. Showing employees that 45+ AI tools are being used across the organization, with specific data types at risk, is more persuasive than abstract security training.


    Audit Cadence

    Shadow AI is not a one-time problem. New AI tools launch weekly, employee adoption patterns shift, and the boundary between sanctioned and unsanctioned blurs as teams experiment.

    Run the full audit quarterly. Between full audits, maintain continuous monitoring via network logs and DLP alerts. Update the AI service domain list monthly.

    For the broader context on why shadow AI is a $19.5M risk and the structural fix beyond auditing, see Shadow AI: The $19.5M Enterprise Risk Your Security Team Can't See.

    Turn unstructured data into AI-ready datasets — without it leaving the building.

    On-premise data preparation with full audit trail. No data egress. No fragmented toolchains. EU AI Act Article 30 compliance built in.

    Keep reading