AI Policy Desk · Governance

Shadow AI: What It Is and How to Prevent It

Shadow AI is the use of unauthorized AI tools inside your company. Learn what it is, why it happens, and six practical steps to prevent it.

Back to blog

Shadow AI: What It Is and How to Prevent It

Shadow AI is the use of AI tools inside your company without approval, IT visibility, or governance controls. Employees paste customer data into ChatGPT, use AI browser extensions on internal documents, or run code through Copilot — none of it tracked, none of it approved.

It happens in every team. And in small teams, the risk is higher because there is rarely anyone watching.


Why Shadow AI Happens

Shadow AI is not malicious. Employees use unauthorized AI tools for the same reason they always adopted unsanctioned software: the tools are faster, more useful, and easier than the approved alternatives — or there are no approved alternatives at all.

According to a 2024 Microsoft survey, more than 75% of knowledge workers already use AI at work, and a significant portion use tools not provided by their employer.

The gap between "what people use" and "what IT knows about" is shadow AI.


What Gets Exposed

When an employee pastes text into an external AI tool, that text leaves your environment. Depending on the tool and account type:

Data type Common example Risk
Customer PII Names, emails in a support summary GDPR/CCPA violation
Internal strategy Product roadmap pasted for a summary Competitive leak
Code and credentials API key in a code snippet Security breach
Legal documents Contract clauses for plain-English translation Privilege waiver risk
Financial data Revenue figures for a report draft Insider information

The AI vendor may use free-tier prompts for training. Even if they do not, the data has left your control.


How to Detect Shadow AI in Your Team

You do not need enterprise security tooling. Start with these three steps:

  1. Run a team survey. Ask: "What AI tools are you currently using for work?" Anonymous surveys get more honest answers. Most people will tell you.
  2. Check browser extensions. AI writing assistants, grammar tools, and screenshot tools often have AI backends. A five-minute scan of what is installed on company devices reveals a lot.
  3. Look at network traffic. If you have basic DNS logging (most business routers include this), search for openai.com, anthropic.com, gemini.google.com, and similar. You will see usage patterns without reading content.

Six Steps to Prevent Shadow AI

These steps work for a team of 5 or 50. They do not require a compliance team.

  1. Publish an approved-tools list. A one-page doc listing which AI tools are allowed, for what purposes, and under what conditions. If employees do not know what is approved, they will guess — or not bother asking.

  2. Write a one-page AI usage policy. Cover three things: what data must never go into an AI tool (customer PII, credentials, confidential documents), what the approval process is for a new tool, and who to contact with questions. Download our AI Policy Template to get started in 30 minutes.

  3. Give employees an approved alternative. Shadow AI spikes when there is no sanctioned option. If your team wants AI writing help, give them a company ChatGPT Team account or equivalent. Friction is the enemy of compliance.

  4. Add AI tools to your onboarding checklist. New employees should learn your AI policy in week one — the same way they learn your password policy. This sets expectations before habits form.

  5. Run a quarterly usage review. Ask teams to share what AI tools they are using and what for. No punishment, no surveillance — just awareness. This gives you early warning before a policy violation becomes a data incident.

  6. Create a fast-track approval path. If requesting a new AI tool takes two weeks and three signatures, employees will skip the process. A simple one-page approval form with a 48-hour turnaround removes the incentive to go rogue.


What Not to Do


Quick-Start Checklist

Copy this into a doc and assign an owner:


Next Steps

Shadow AI is a symptom of a governance gap, not a people problem. Close the gap with a clear policy and approved tools, and the shadow disappears.