Your staff are already using AI. The question is whether you know about it — and whether you know where their conversations are going.
Right now, somebody in your organisation is pasting client data into ChatGPT. Someone else is summarising a confidential report with Claude. A third person is using Gemini to draft a proposal with commercially sensitive figures. They are not being reckless. They are being productive. But every one of those conversations is being stored, processed, and — on most plans — used to train the next generation of AI models.
This is not a theoretical risk. It is happening today, across every sector, in businesses of every size. And the data policies of the three largest AI providers are far less protective than most people assume.
The good news? You do not have to choose between AI productivity and data safety. You can have both. Here is how.
shadow AI usage
per breach from shadow AI
when training is enabled
The Shadow AI Problem
Shadow AI is the term for employees using AI tools without IT knowledge or approval. It is not new — shadow IT has been a headache for decades — but AI has made it exponentially more dangerous. When someone uses an unauthorised spreadsheet, the risk is limited. When someone pastes a client contract into a public AI tool, that data enters a training pipeline you have zero control over.
Research from IBM’s 2026 Cost of Data Breach Report found that 76% of UK organisations have shadow AI as a definite or probable challenge. The same report found that shadow AI incidents add approximately £670,000 to the average cost of a data breach. That is not the total cost — that is the additional cost on top of an already expensive incident.
Why it happens
Most shadow AI usage comes from well-intentioned employees trying to work faster. They sign up for free ChatGPT accounts, paste in client documents for summarisation, use Claude to draft emails, or feed Gemini with financial data for analysis. They do not read the terms of service. They do not think about where the data goes. The answer is not to ban AI — it is to give your team a better way to use it.
OpenAI (ChatGPT): What Happens to Your Data
OpenAI’s data policies depend entirely on which plan you are using. The differences are significant — and the default settings on consumer plans should concern any business handling sensitive data.
Consumer plans: Free, Plus, and Pro
Training: ChatGPT trains on your conversations by default. There is an opt-out toggle buried in settings, but enabling it disables your chat history. Most users never find it, and even fewer use it.
Retention: Conversations on consumer plans are retained indefinitely. OpenAI was court-ordered to preserve all chat logs from May to September 2025 as part of the New York Times copyright lawsuit — a reminder that your conversations can become legal evidence.
Data residency: UK data residency is available only for Enterprise customers. Consumer plan data is processed and stored in the United States.
Advertising: As of February 2026, OpenAI is running advertisements on the free tier. Your conversations now exist alongside an ad-supported business model.
Business plans: Team, Enterprise, and API
Training: OpenAI does not train on data from Team, Enterprise, or API usage. This is a contractual guarantee backed by a Data Processing Agreement.
Retention: Admin-controlled retention policies. Enterprise customers can set their own data lifecycle rules.
OpenAI certifications
OpenAI holds ISO 27001 and SOC 2 Type 2 certifications for its enterprise offerings. These are meaningful security baselines, but they do not change the fundamental data handling on consumer plans. The models themselves — GPT-4o, o3, o4-mini — are exceptionally capable. The problem is not the intelligence. It is where the data goes.
Anthropic (Claude): What Happens to Your Data
Anthropic has positioned Claude as the “safety-first” AI company. But a significant privacy policy change in September 2025 and two data leaks in early 2026 have complicated that narrative.
Consumer plans: Free, Pro, and Max
Training: Since September 2025, Anthropic trains on Free, Pro, and Max conversations by default. This was a notable pivot — prior to that date, Anthropic did not train on user conversations. The change was buried in a terms of service update that most users did not notice.
Retention: When training is enabled, Anthropic retains conversation data for up to 5 years. That is one of the longest retention periods of any major AI provider.
Data residency: All Claude data is stored in the United States. There is no UK or EU data residency option for any Anthropic plan. The only workaround is accessing Claude through AWS Bedrock in an EU region, but this requires technical implementation and is not available to most business users.
Business plans: Work, Enterprise, and API
Training: Claude for Work, Enterprise, and API tiers do not train on your data. API retention is limited to 7 days for trust and safety purposes.
Recent incidents
In March 2026, Anthropic experienced two separate data exposure events: one involving the Mythos model’s training data and another involving Claude Code source material. Additionally, Anthropic reached a $1.5 billion copyright settlement. Despite these incidents, Claude remains one of the most intelligent models available — Opus 4 is widely considered the strongest reasoning model on the market. Again, the issue is not capability. It is data control.
Google (Gemini): What Happens to Your Data
Google’s Gemini has the broadest distribution of any AI assistant — it is embedded in Gmail, Docs, Search, and Android. That reach makes its data policies particularly important to understand.
Consumer plans: Free Gemini
Training: Google trains on free Gemini conversations by default. There is no ambiguity here — it is stated clearly in their terms.
Retention: Conversations on the free tier are subject to a 3-year human review retention period. That means Google employees can read your conversations for quality assurance and model improvement for up to three years after you send them.
Business plans: Workspace and Enterprise
Training: Google Workspace plans include a contractual guarantee that Gemini does not train on your data. This is backed by a Data Processing Agreement (DPA) — something the consumer tier does not offer.
Data residency: UK data residency has been available since February 2026 for Enterprise and Workspace customers. This is a genuine advantage over Anthropic, which offers no European residency at all.
GeminiJack: the zero-click exploit
In June 2025, security researchers disclosed “GeminiJack” — a zero-click exploit that allowed attackers to exfiltrate corporate data through Gemini without any user interaction. The attack leveraged Gemini’s deep integration with Google Workspace to extract sensitive documents. Google patched the vulnerability, but the incident demonstrated the risks of tight AI-productivity suite integration.
Certifications: Google holds ISO 27001, ISO 42001, SOC 2, and FedRAMP certifications — the broadest certification portfolio of the three providers.
Side-by-Side Comparison
Here is how the three providers stack up on the policies that matter most to UK businesses.
| ChatGPT (OpenAI) | Claude (Anthropic) | Gemini (Google) | |
|---|---|---|---|
| Training (Free) | Yes — opt-out toggle | Yes — since Sept 2025 | Yes — by default |
| Training (Enterprise) | No — contractual | No — contractual | No — contractual + DPA |
| UK Data Residency | Enterprise only | None — US only | Enterprise + Workspace |
| Retention (Free) | Indefinite | 5 years (if training on) | 3-year human review |
| Retention (Enterprise) | Admin-controlled | 7-day API / admin-set | Admin-controlled |
| Recent Incidents | NYT lawsuit log preservation; ads on free tier | Two March 2026 data leaks; $1.5B copyright settlement | GeminiJack zero-click exploit; Personal Intelligence class-action |
The pattern is clear: every provider is safe at the enterprise tier and unsafe at the consumer tier. The problem is that your employees are not using the enterprise tier. They are using free accounts with their personal email addresses.
But here is the thing most articles miss: these are brilliant models. ChatGPT, Claude, and Gemini represent the most capable AI ever built. The intelligence is not the problem. The data handling is. And that is a solvable problem.
How Nerdster.ai Solves This
At Nerdster.ai, we believe you should not have to choose between AI capability and data safety. Every major AI model — GPT-4o, Claude Opus, Gemini Pro, Llama, Mistral — is available through secure APIs and private hosting. The question is not which model to use. It is how to use it without exposing your clients’ data.
That is exactly what we do. We work with UK professional services firms to deploy AI that is as intelligent as anything you can get from ChatGPT or Claude — but where your data never leaves your control.
We help you choose and use the best AI models
Not every task needs the same model. A contract review might need Claude’s deep reasoning. A quick email draft might work best with GPT-4o. A coding task might suit an open-source model running locally. We help you understand which model fits which workload — and set up the secure infrastructure to use them all without worrying about data exposure.
- Model selection: We evaluate ChatGPT, Claude, Gemini, Llama, Mistral, and other leading models against your specific use cases. You get the most intelligent model for each task — not a one-size-fits-all compromise.
- Secure API access: We configure enterprise-tier API access with Data Processing Agreements, admin-controlled retention, and full audit trails. Your team gets the same AI capability with proper contractual protections.
- Training and onboarding: We train your staff on how to use AI effectively and safely. No more shadow AI — because the approved tools are better than the free alternatives they were using anyway.
We build custom AI agents for your workloads
Generic AI tools are powerful but inefficient. They do not know your processes, your templates, your compliance requirements, or your clients. A custom AI agent does.
- Document review agents: Trained on your firm’s templates and regulatory requirements. They review contracts, flag risks, and summarise key terms — in seconds, not hours.
- Client intake agents: Automate the initial information gathering, conflict checks, and matter setup. Your fee earners spend time on billable work, not admin.
- Research agents: Search across your internal knowledge base, precedents, and public sources simultaneously. Every response is cited and auditable.
- Reporting agents: Generate client reports, management information, and compliance documentation automatically from your existing data.
These agents run on the most capable AI models available — but they are configured to work within your data boundaries. No client information leaves your environment. Every interaction is logged. Every output is traceable.
We offer your own private LLM — locally or privately hosted
For firms that need absolute data control, we deploy private large language models that run entirely on your infrastructure. This is not a watered-down version of AI. Modern open-source models like Llama 4 and Mistral Large rival the performance of ChatGPT and Claude on most business tasks — and they run on hardware you own.
What private hosting looks like
On-premises: We install and configure AI models on your existing servers or dedicated GPU hardware in your office. The models run on your network. Nothing touches the internet. Ideal for law firms, financial advisers, and healthcare providers handling the most sensitive client data.
Private cloud
Dedicated cloud: We deploy AI models in your own AWS, Azure, or Google Cloud tenancy — in the UK region of your choice. You get the scalability of cloud computing with the data isolation of on-premises. All traffic stays within your virtual private cloud. No shared infrastructure. No third-party data access.
Air-gapped deployment
Air-gapped: For maximum security, we deploy air-gapped AI systems on hardware with no internet connection whatsoever. The AI processes your data in a sealed environment. No network traffic in or out. This is the gold standard for firms handling privileged legal matter, financial due diligence, or classified information.
With any of these options, you get the full capability of modern AI — document analysis, summarisation, drafting, coding, research, translation — without a single byte of client data leaving your control. Every query and every response is logged within your systems, giving you a complete audit trail for regulators.
What This Means for Your Firm
If you operate in a regulated sector — legal, financial services, healthcare, accounting — the implications of public AI data policies are severe. But the solution is straightforward.
The regulatory landscape
- SRA (Solicitors Regulation Authority): The SRA expects law firms to demonstrate appropriate AI governance. Private AI with full audit trails meets this standard. Consumer ChatGPT does not.
- FCA (Financial Conduct Authority): The FCA applies existing frameworks — including Consumer Duty and SM&CR individual accountability — to AI usage. Custom AI agents with built-in compliance checks help your advisers meet these obligations.
- EU AI Act: The majority of provisions come into force on 2 August 2026 — less than four months from now. Private AI deployments give you the transparency, auditability, and human oversight the Act requires.
“The risk is not that AI is dangerous. The risk is that uncontrolled AI creates liability you cannot manage, audit, or defend to a regulator. Private AI eliminates that risk entirely.”
Getting Started with Nerdster.ai
You do not need to solve this overnight. But you do need to start. Here is how working with us typically looks.
1. Free AI audit (30 minutes)
We identify which AI tools your staff are using, what data is being shared, and where your compliance gaps are. You get a clear picture of your current shadow AI risk — and a practical roadmap for fixing it.
2. Model selection and agent design
We evaluate your workflows and recommend which AI models and custom agents will deliver the most value. Contract review? Client intake? Research? Reporting? We design agents tailored to how your team actually works — not generic chatbots.
3. Secure deployment and training
We deploy your AI solution — whether that is secure API access, custom agents, a private LLM, or an air-gapped system — and train your team to use it. The result: your staff stop using shadow AI because the approved alternative is faster, smarter, and easier. Shadow AI disappears because you have replaced it with something better.
The Bottom Line
ChatGPT, Claude, and Gemini are extraordinary tools. They are also, on their consumer plans, extraordinary risks for any business that handles client data. The default settings train on your conversations. The retention periods are measured in years. The data residency is overwhelmingly US-based. And none of them offer a Data Processing Agreement on free or personal plans.
But the intelligence behind these models is available to you — privately, securely, and on your terms. You can use the most capable AI ever built without worrying about where your client data ends up. You can build custom AI agents that know your business and your compliance requirements. You can run your own private LLM that never connects to the internet.
That is what Nerdster.ai does. We help UK businesses understand, choose, and deploy AI safely. If that sounds like what your firm needs, let’s talk.