Sequentur Blog
Helping you stay ahead of IT challenges
Real-world IT knowledge from engineers solving problems every day.
Practical IT knowledge for businesses that can’t afford downtime
Your employees are already using AI at work: what that means for your business
Most small business owners think AI adoption is a future decision. It is not. The decision has already been made – just not by leadership. It was made by the marketing coordinator who pastes draft emails into ChatGPT to clean up the tone. By the bookkeeper who uses Gemini to explain a confusing IRS notice. By the salesperson who feeds a customer’s contract into Claude to summarize the terms. By the office manager who has Copilot reading the entire SharePoint library to write meeting prep notes. By the engineer who uploads code snippets to GitHub Copilot. By the HR director who runs candidate resumes through ChatGPT to draft interview questions.
None of these people are doing anything malicious. They are doing what employees have always done with new technology – using the tool that makes their job easier, before anyone tells them whether they are allowed to. The category name for this is shadow AI, and it is the IT, security, and compliance problem most small businesses have not yet realized they have.
This article is the wake-up call. It explains what shadow AI is, what data your employees are very likely already feeding into AI tools, what the consumer versions of those tools do with that data, why the gap between what leadership thinks is happening and what is actually happening is wider than you expect, and why this is an IT and compliance issue – not just an HR one. If you are an SMB owner, an operations manager, an HR director, or the in-house IT generalist trying to get ahead of this, start here.
Short answer
Your employees are using ChatGPT, Microsoft Copilot, Google Gemini, Claude, and a long tail of niche AI tools – some sanctioned, most not. The free and consumer versions of these tools train on user inputs by default, retain data for varying periods, and sit outside your identity, logging, and data-protection controls. Sensitive material is already going in: client lists, financial data, HR records, source code, contract language, and confidential strategy. Surveys consistently show 60-75% of knowledge workers use AI at work, but only 20-30% of businesses have an AI policy in place. The gap is the risk. Closing it does not require banning AI – it requires an acceptable use policy, an approved-tools list, the right enterprise tier where data sensitivity demands it, and a short, non-judgmental conversation with staff. The fastest first move is to assume shadow AI is already happening in your business and act accordingly.
Shadow AI at a glance
| Question | Short answer |
|---|---|
| What is shadow AI? | Employees using AI tools without IT approval, logging, or governance |
| How common is it? | 60-75% of knowledge workers use AI at work; most are not sanctioned |
| What data is going in? | Client info, financials, HR records, source code, contracts, internal strategy |
| Why does it matter? | Consumer AI tools train on inputs by default; data lives outside your controls |
| Is it an HR problem? | No – it is IT, security, compliance, and HR. All four. |
| Can you block it? | Partially. Not fully. And blocking without offering an approved alternative makes shadow use worse. |
| What is the cheapest first step? | Write a one-page AI acceptable use policy and tell staff what is approved |
| When does this become a compliance problem? | Immediately, if you handle PHI, PII, regulated financial data, or contractual confidentiality obligations |
| How long does an initial AI governance setup take? | 2-4 weeks for a small business – policy, approved-tools list, staff communication, basic logging |
| What is the biggest mistake? | Assuming it is not happening because no one has asked permission |
What shadow AI actually means in an SMB
Shadow AI is a specific instance of the older “shadow IT” problem – employees using technology that IT does not know about, has not approved, and cannot see into. Dropbox accounts created with personal email addresses. Slack workspaces spun up for “just this one project.” Personal Google Drives used to share client files because the corporate file share is awkward to access from home. The pattern is the same every time: the employee has a job to do, the sanctioned tool is missing or friction-heavy, and a consumer alternative is one click away.
Shadow AI is that exact pattern, with one important difference. The consumer versions of AI tools are unusually good at the job. ChatGPT actually does write a passable first draft of a marketing email. Copilot really does summarize a 40-page contract in a way that saves an hour. Gemini genuinely does help an analyst clean up a messy spreadsheet. The productivity gain is real and immediate, which means the incentive to use the tool is much stronger than the incentive to use, say, an unsanctioned file-sharing service. Employees who would not casually upload a customer database to Dropbox will paste a customer database into ChatGPT without a second thought, because the task they are doing feels like writing, not file sharing.
The forms shadow AI takes in a typical SMB:
- Personal-account AI use on personal devices for work tasks. The bookkeeper uses her home laptop in the evening to ask ChatGPT to explain a regulatory notice. The query includes the client name.
- Personal-account AI use on work devices. The marketing coordinator signs into ChatGPT with a personal Google account on his work laptop. Everything he pastes – draft emails, customer lists, campaign performance data – sits in his personal ChatGPT history forever.
- Browser extensions and AI-powered plugins. A “summarize this page” Chrome extension is sending the contents of every page the employee views (including the CRM, the M365 admin center, payroll dashboards) to a third-party AI vendor for processing.
- AI features baked into approved SaaS tools the business never reviewed. Notion AI, Slack AI, HubSpot Breeze, Zoom AI Companion, Otter.ai, Fireflies, Read.ai. The business approved Slack. It never approved Slack AI, because Slack AI did not exist at the time. It is on now.
- Free consumer AI tools used on confidential data because no enterprise alternative was provided. This is the most common pattern. Staff want to use AI. There is no approved option. They use the free one and feel guilty but productive.
The unifying property: leadership does not know it is happening, IT cannot see it, no one has classified what data should and should not be going in, and there is no contractual agreement with the AI vendor that protects the business.
What data your employees are very likely feeding into AI tools
A 2024-2025 spread of independent surveys – from Cyberhaven, LayerX, Microsoft, Gartner, and others – converges on a consistent picture. Employees paste highly sensitive material into AI tools all the time. They do it because the tool is helpful, because they are not thinking of “pasted text” as “transmitted data,” and because most companies have never told them not to. The categories show up across every survey:
- Customer and client data. Names, contact information, order histories, account numbers, support ticket contents. Pasted into AI tools to draft personalized outreach, summarize cases, or analyze patterns. The customer never consented to having their information processed by a third-party AI vendor.
- Financial information. Bank statements, P&L spreadsheets, invoices, tax notices, payroll runs. Pasted into AI tools to interpret a number, draft a response to a vendor, or “ask ChatGPT what this means.”
- HR records. Employee names, salaries, performance reviews, candidate resumes, disciplinary notes, immigration paperwork. Pasted into AI tools to write reviews, sort candidates, or draft hard conversations. Personal data on employees is regulated under multiple frameworks (state privacy laws, GDPR if any EU employees, sector-specific rules).
- Source code and proprietary algorithms. Engineers paste production code into AI for review, debugging help, or to generate tests. The code may contain credentials, business logic that is competitively sensitive, or copyright-bound content. Samsung famously banned ChatGPT internally in 2023 after engineers pasted source code from semiconductor designs – the same risk applies at any scale.
- Legal documents and contracts. Draft contracts, NDAs, master service agreements, client engagement letters. Pasted in for summary, redlining, or interpretation. Many of these contain confidentiality clauses the employee just violated by pasting them.
- Health information (PHI). In healthcare-adjacent SMBs and businesses with even incidental health data (life insurance, employee benefits administration, fitness, wellness apps), PHI ends up in AI tools. This is a HIPAA violation the moment it happens.
- Internal strategy, plans, and decks. Board materials, pricing strategy, acquisition targets, layoff plans. Pasted in to clean up the prose. Now lives in a third party’s data store.
- Credentials and secrets. API keys, database connection strings, internal URLs, sometimes passwords – employees paste config files or stack traces in for debugging help, not realizing what is in them.
If the response is “our people would not do that” – they would, and they are, statistically. The same surveys that find this pattern find that the offenders are not junior or careless; they are the most productive employees, who have figured out that AI makes them faster.
Why most free AI tools train on your inputs by default
Here is the part that catches business owners off guard. The free consumer versions of the major AI tools have, historically, used user inputs to train future versions of the model by default. The terms have shifted over time and vary by vendor, but the pattern is consistent enough to be worth understanding.
- ChatGPT free and Plus tiers. OpenAI’s default setting for the consumer ChatGPT product has historically been to use conversations for training. Users can opt out, but most do not, and many do not realize the setting exists. The Enterprise, Business (Team), and API tiers do not train on inputs by default.
- Google Gemini consumer. Google’s consumer Gemini product retains conversations for a configurable period and uses human reviewers to improve the model. Gemini for Google Workspace operates under different terms with stricter data controls.
- Microsoft Copilot consumer (formerly Bing Chat). The free Copilot product operates under consumer terms. Copilot for Microsoft 365 – the licensed product that integrates with Outlook, Word, Excel, and Teams – operates under enterprise data protection terms that explicitly exclude training on customer data.
- Anthropic Claude consumer. Claude’s consumer product also has training opt-in/opt-out controls that vary by region. Claude for Work and the Anthropic API have different default behaviors that lean toward data protection.
The general rule of thumb: if you are not paying for the enterprise or business tier, assume the vendor reserves more rights over the data than you would want. “Free” AI is paid for, in part, with your inputs.
This matters for two reasons. First, the data is leaving your business. Second, depending on the vendor and the moment, the data may be used to train a future model – which means a future version of the AI could, in theory, surface variants of your confidential information when prompted by someone else. The actual likelihood of verbatim re-emission is low for well-trained models, but the legal and contractual problem is not the technical likelihood – it is the fact that you, as the data controller, have moved confidential information to a third party under terms you never reviewed and a contract you never signed.
If you handle anything regulated – PHI under HIPAA, financial data under GLBA, payment card data under PCI, EU resident data under GDPR, California resident data under CCPA, or contractually confidential information under client agreements – this is not theoretical. It is a compliance gap that exists in your business right now.
The gap between what leadership thinks and what is actually happening
This is the part of the problem that resists belief. Most owners assume that because no one has asked for an AI tool, no one is using one. The data says otherwise.
Across recent industry surveys, the same shape repeats:
- 60-75% of knowledge workers report using AI at work, depending on which survey you read
- Only 20-30% of those workers say their employer has provided an AI tool
- Roughly half of users do not tell their manager they are using AI
- A meaningful minority – 20-40% in some surveys – explicitly use AI on tasks their employer would not approve, knowing it would not be approved
Translated to a 20-person SMB: at least 12 of those 20 people are using AI for work today. Eight to ten of them are using consumer tools you did not approve. Four to six are doing so on data that creates real exposure for the business. Two or three are doing it on something that, if a regulator or a client auditor saw, would create a notification obligation.
This is not because the staff are reckless. It is because the staff are productive, and the productivity tools the employer has provided do not include AI, and the productivity tools the open internet provides are AI. The mismatch is the problem.
The other half of the gap is what leadership thinks is the worst-case scenario. The instinct is to imagine an employee pasting the customer database into ChatGPT – a dramatic, single-event leak. The actual pattern is much less dramatic and much more compounding. Twenty employees, each pasting a slightly sensitive thing into a free AI tool two or three times a week, for two years. The data lives somewhere outside the business. The business has no record of what was sent, when, by whom, or to which vendor. When the question finally arrives – from a client during a security questionnaire, from an insurer during a renewal, from a regulator during an audit – there is no answer to give except “we do not know.”
Why this is an IT and compliance issue, not just an HR one
When shadow AI first becomes visible inside a business, the instinct is to treat it as an HR problem. Someone misbehaved. Someone needs to be reminded of the rules. The HR-only frame is wrong for four reasons.
1. It is a data control issue. Once data leaves the business and enters a third-party AI vendor’s systems, it is governed by that vendor’s terms. The business cannot recall it, cannot guarantee its deletion, and cannot make any representation about it to clients, regulators, or auditors. Data control is IT and security, not HR.
2. It is a contractual exposure. Client contracts, vendor agreements, NDAs, and master service agreements routinely contain confidentiality clauses that prohibit sharing client information with third parties without consent. Pasting client information into a consumer AI tool is “sharing with a third party,” whether or not the employee thought of it that way. The business is the contracting party, not the employee. Legal and operations own this, not HR.
3. It is a regulatory exposure. HIPAA requires Business Associate Agreements with anyone who processes PHI. PCI requires controls on where cardholder data flows. State privacy laws and GDPR require known data flows and the ability to honor deletion requests. Consumer AI tools sit outside all of these. The moment regulated data lands in one, the business has a compliance gap that did not exist five minutes earlier.
4. It is a cybersecurity issue. AI tools are now an exfiltration vector. Some are deliberate (employees uploading data to a competitor’s AI to compare offers). Most are accidental. Either way, the question “what data has left our business this quarter, and where did it go” is a security question, and shadow AI is the channel most SMBs cannot answer for.
The framing matters because the response is different. An HR response is policy plus a stern email plus a training video. The required response is a policy plus an approved-tools list plus a Microsoft 365 enterprise tier where Copilot is appropriate plus a data classification rule plus visibility into AI tool use plus an incident response procedure for AI-related leaks. HR owns the policy and the communication. IT owns the tooling. Operations owns the data classification. Legal owns the contractual review. Whoever owns compliance – in many SMBs that is one person wearing four hats – has to make sure the whole thing actually hangs together.
What the first move looks like
The instinct, when an owner realizes shadow AI is happening, is one of two extremes: ban AI entirely, or pretend it is fine because the productivity is real. Neither extreme works.
Banning AI entirely does not stop usage; it just pushes it further into the shadows. Staff who were using AI on a work laptop with a personal account will switch to using AI on a personal device, where the business has even less visibility. The same data still leaves the business; you just see it less.
Pretending it is fine because the productivity is real ignores the compliance and contractual exposure, which compounds quietly until a question shows up that you cannot answer.
The middle path is the only one that works: assume staff want AI, give them a sanctioned way to use it, and write down what is allowed and what is not. The components are:
- A short AI acceptable use policy. One to three pages. What is approved, what is forbidden, what data categories must never go into any AI tool, and what to do if you think something sensitive went in by mistake. The full breakdown – what each section should cover, plus a sample one- to three-page outline you can adapt – is in how to write an AI acceptable use policy for your small business. It builds on the same foundations as a general cybersecurity policy for small business and a remote-team IT policy – both of which most SMBs already need anyway.
- An approved-tools list. Usually a short list. For most SMBs running Microsoft 365, the answer is Copilot for Microsoft 365 with the licensing that includes enterprise data protection – which assumes the underlying tenant has been hardened first (see Microsoft 365 security hardening for small business). For non-M365 shops, the answer is usually ChatGPT Team or Enterprise, or Claude for Work, or Gemini for Google Workspace. There can also be sanctioned use of consumer free tools for non-confidential tasks – that needs to be explicit, not assumed.
- A data classification rule, even a simple one. Three tiers is enough: public/general, internal, and sensitive (covering client data, financials, HR, regulated information, and confidential strategy). The rule for AI use is: public and internal data can go into approved AI tools; sensitive data only in tools with enterprise data protection contractually in place; never in free or unapproved tools.
- A short staff conversation. Non-punitive. The framing is “we know AI is useful, we are not banning it, we want to give you the right tools and make sure the wrong data does not end up in the wrong place.” Most staff will be relieved. The shadow part of shadow AI is not defiance; it is the absence of a sanctioned path.
- A BYOD position. Personal devices are the hardest part of the AI governance problem because the business has the least control over them. A clear position on whether personal devices can be used for work AI, and on what conditions, has to be in the BYOD policy.
Where this stops being a DIY job and starts being a managed-services job depends on the size of the business and the data sensitivity. A 5-person business with no regulated data can run an AI acceptable use policy on its own. A 30-person business handling client financial information, healthcare data, or contractually confidential material has more moving parts than the in-house generalist can keep current as the AI landscape changes monthly. The signal that DIY has reached its ceiling is the same as for the rest of IT – see the patterns in signs your small business has outgrown DIY IT.
What this costs to ignore
The downside of not getting ahead of shadow AI sits in a few specific buckets. They are easy to dismiss individually and hard to dismiss in combination.
Breach disclosure obligations. If sensitive personal data goes into a consumer AI tool, depending on the data type and jurisdiction, it may meet the legal definition of a breach. State data breach notification laws, HIPAA, GDPR, and CCPA all have notification thresholds that AI tool leaks can cross. The average notification event for a small business runs into five and six figures once forensics, legal, and customer notification are accounted for – see how much does a data breach cost a small business for the underlying math.
Contractual breach with clients. If a client contract includes a confidentiality clause (most do), and confidential client information ended up in a consumer AI tool, the business has breached the contract. The exposure depends on the contract but typically includes indemnification obligations, potential termination, and reputational damage that survives the legal outcome.
Insurance complications. Cyber insurance policies increasingly include questions about AI governance during renewal. Insurers want to know what AI tools are approved, what controls are in place, and what data classification rules apply. Answers like “we have not addressed this” can affect renewal terms, premium, or coverage scope.
Regulatory exposure. For SMBs in regulated verticals – healthcare, financial services, defense contracting, legal services – shadow AI sits inside a regulator’s blast radius. HIPAA, GLBA, FTC Safeguards, DFARS / CMMC, and state-level AI laws all touch this. Audit findings citing “ungoverned AI tool use” are increasingly common.
Competitive exposure. Less measurable but real. Strategy decks, pricing analysis, and product plans pasted into consumer AI may, in extreme cases, leak through training. More commonly, they sit in the vendor’s logs – a single subpoena, vendor breach, or vendor policy change can change what happens to them.
None of these are imminent for every business on every Tuesday. Collectively, they are why “we have not addressed this” stops being a defensible answer somewhere between the 10-person and 30-person mark, and stops being a defensible answer immediately if any regulated or confidential data is in scope.
Ten common shadow AI mistakes SMBs make
- Assuming no one is using AI because no one has asked. The data says they are. Assume yes, then act accordingly.
- Banning AI as the first response. Drives use further into the shadows, removes the leverage of being able to approve specific tools.
- Treating it as an HR problem only. It is data control, contracts, compliance, security, and culture. HR is one of five owners, not the only one.
- Approving Copilot for M365 without hardening the tenant first. Copilot sees everything the user can see. Overpermissioned users mean Copilot becomes a data leakage tool internally even before external concerns.
- Buying enterprise AI licensing without a data classification rule. The license protects the data the employee remembers to put in the right place. Without a classification rule, employees still paste sensitive data into the wrong tool.
- Writing an AI policy and never communicating it. A policy filed in SharePoint that no one has seen is worse than no policy – it creates the illusion of governance without the substance.
- Forgetting browser extensions and SaaS-embedded AI. The marketing automation tool, the meeting recorder, the CRM, the file storage app – all may have added AI features that send your data somewhere new since the last contract review.
- Letting the policy go stale. AI vendor terms change every quarter. Tools that were not safe last year are safe now, and vice versa. A policy reviewed once and never updated drifts out of reality within months.
- Ignoring personal-device use. Most shadow AI runs on personal devices. A BYOD-shaped policy that does not address AI specifically misses most of the actual surface area.
- Skipping the staff conversation. Policies land better when employees were involved or at least informed before the policy went live. Surprise policies create resentment without changing behavior.
How long it takes to get this from “we have not addressed it” to “we have a handle on it”
For a typical SMB that is starting cold, the realistic timeline is two to four weeks of calendar time, with maybe 15-25 hours of total effort spread across the owner, IT, and HR.
| Phase | What gets done | Typical duration |
|---|---|---|
| Discovery | List the AI tools you know are in use, ask staff what they actually use, review SaaS contracts for AI features | 1 week |
| Classification | Define data tiers (public, internal, sensitive), map what data sits in which tier | 2-3 days |
| Policy draft | Write the 1-3 page AI acceptable use policy, including approved-tools list and breach reporting | 3-5 days |
| Licensing decisions | Decide whether to license Copilot for M365, ChatGPT Team, etc. – and for which staff | 1 week (often in parallel with policy) |
| Tenant prep | Tighten M365 permissions before turning Copilot on (sensitivity labels, conditional access, file sharing review) | 1-2 weeks |
| Communication | Staff conversation, policy publication, training | 2-3 days |
| Operate | Ongoing – quarterly policy review, periodic spot checks, update as vendors and laws change | Permanent |
For a regulated business (healthcare, financial services, legal, defense), add another week to two weeks for the compliance overlay – HIPAA BAA review with AI vendors, GLBA Safeguards mapping, CMMC controls cross-walk where relevant. This is also where the work usually stops being a sole-owner project and starts needing either a competent in-house generalist with bandwidth, or external help.
What is next in this content series
This article is the wake-up call. The follow-ups go deeper into each component:
- A practical AI acceptable use policy template – the next article in this series
- What data is and is not safe to put into specific AI tools, vendor by vendor
- AI and HIPAA, including which AI vendors have signed BAAs and what that actually covers
- The AI security risks every small business should know about (prompt injection, data leakage, AI-powered phishing, deepfakes, voice cloning)
- Microsoft Copilot for small business specifically, including the M365 permission hygiene that has to happen before rollout
- How to build a lightweight AI governance framework that does not require enterprise-scale process
If your AI governance work is sitting inside a broader managed cybersecurity engagement, the relevant parent context is the managed cybersecurity services for small business overview. If the business is in a regulated vertical, the HIPAA cybersecurity requirements article covers the foundation that AI governance has to sit on top of.
How Sequentur can help
If you suspect shadow AI is already happening in your business and want help getting a policy, an approved-tools list, and the right Microsoft 365 configuration in place – or just a second pair of eyes on what you have already drafted – schedule a call.
Get the Best IT Support
Schedule a 15-minute call to see if we’re the right partner for your success.
Testimonials
What Our Clients Say
Here is why you are going to love working with Sequentur