How Slack, LinkedIn, Anthropic and Others Use Your Work Data for AI
Freelance platforms are just one piece of your company's data footprint. A typical knowledge-work company also routes sensitive information through:
Each of these has its own AI training policy, data retention rules, and opt-in/opt-out mechanics. And they're evolving fast—what was "safe" six months ago may have changed.
This guide maps the current landscape across platforms:
Your company pays for Slack Enterprise and Google Workspace (protected). But employees also use:
These personal/free accounts don't have enterprise protections. Company data is leaking into consumer-tier training sets because employees don't realize the distinction.
How it works: Platform uses aggregate, de-identified data to train traditional machine learning models (recommendation engines, search ranking, spam detection). Individual messages or files aren't directly accessed by the models.
Example: Slack's Traditional ML (Pre-AI Features)
Before generative AI features, Slack used de-identified, aggregate data for search relevance and channel recommendations. Their April 2025 clarification states:
"Machine learning models are trained on de-identified and aggregated data. They do not access the content of customer messages in DMs, private channels or public channels."
Risk level: Low. De-identification and aggregation reduce exposure of specific confidential content. But "what gets discussed" patterns (topics, frequency, user behavior) still inform platform improvements.
How it works: Platform offers generative AI features (chatbots, code completion, writing assistants). To power these, the platform trains on actual user inputs—messages, prompts, code. Users can opt in, opt out, or toggle per-workspace.
Example: Anthropic Claude's 2025 Policy Shift
Anthropic (maker of Claude) initially took an "opt-in" approach: by default, they would not use customer data to train models. In 2025, they shifted:
Risk level: Medium-High for consumer users. If your team uses Claude.ai (free or Pro), be aware that opting in means conversations—including company strategy, client information, draft contracts—may train future models.
Mitigation: Use Claude's API or Teams plan for business work, which excludes training. Or configure opt-out settings for consumer accounts.
How it works: Platform's privacy policy contains vague language ("we may use data to improve services"). Users discover (or fear) that their messages, code, or professional content is being used for AI training. Public backlash or lawsuits force clarification.
Example: LinkedIn Private Messages Controversy (2025)
In 2025, a proposed class action alleged that LinkedIn used private messages of Premium customers to train generative AI models. LinkedIn denied the allegations, produced evidence that private messages weren't used, and the case was voluntarily dismissed.
But the complaint revealed key concerns:
Risk level: Variable. When policies are unclear and litigation is ongoing, companies face uncertainty about whether their historical data has already been used for training and whether current opt-outs are effective.
| Tool | AI Features | Training on User Data? | Opt-Out Available? | Enterprise Carve-Out? |
|---|---|---|---|---|
| Slack (Salesforce) | Slack AI (search, summaries, thread insights) | Traditional ML: De-identified, aggregate data; does not access message content. Generative AI: Uses third-party LLMs (no Slack-trained models on customer data); retrieval-augmented generation only. |
N/A – Slack doesn't train generative models on customer data | ✓ Yes – commercial customers protected by "no customer data used to train LLMs" policy |
| Anthropic Claude | Claude chatbot, API, coding assistance | Consumer (free/Pro/Max): Will train on chats/code sessions if user opts in. Retention up to 5 years for safety classifiers. Commercial/API: Not used for training. |
✓ Yes – consumer users can opt out (must choose after certain date) | ✓ Yes – API and Teams plans exclude training |
| OpenAI ChatGPT | ChatGPT, API, Codex | Consumer (free/Plus): By default, conversations may be used for training. Can opt out via settings. API / Enterprise: Not used for training unless explicitly opted in for fine-tuning. |
✓ Yes – "Data Controls" in settings (consumer); API customers excluded by default | ✓ Yes – API and Enterprise agreements exclude training |
| Google Workspace (Gemini) | Gemini in Gmail, Docs, Sheets; admin AI | Consumer (free Gmail): May use data for training and ads. Workspace (paid): Google states Workspace data (Docs, Gmail, Drive) is not used to train Gemini or ads models. |
Partial – consumer users have limited controls; Workspace customers protected by contract | ✓ Yes – Workspace terms exclude training on customer content |
| Microsoft 365 Copilot | Copilot in Word, Excel, Teams, Outlook | Enterprise: Microsoft states Copilot does not train on customer data; uses retrieval-augmented generation and enterprise-specific grounding. | N/A – not trained on customer data | ✓ Yes – enterprise-only feature, protected by DPA |
| GitHub Copilot | AI code completion, chat | Suggestions: Trained on public GitHub repos. Your private code is not used to train Copilot's base models (as of 2023 policy). Telemetry: Usage data (not code content) may inform product improvements. |
Partial – can disable telemetry sharing; code content not used for training by policy | ✓ Yes – GitHub Enterprise Cloud policies exclude training on private repos |
| Feed recommendations, job matching, learning pathways | Traditional ML: Profile data, connections, activity used for recommendations. Generative AI: Disputed; 2025 lawsuit alleged private messages used for training. LinkedIn denied and case dismissed. "Do not train" setting available but non-retroactive. |
✓ Yes – "Data for Generative AI Improvement" toggle in settings (added 2024-2025) | Unclear – LinkedIn Sales Navigator and Recruiter may have different terms; check enterprise contracts | |
| Notion AI | AI writing assistant, Q&A | Notion states: Customer content is not used to train AI models. AI features use third-party LLMs (OpenAI) under agreements that prohibit training on Notion user data. | N/A – not trained on user data | ✓ Yes – all plans (including free) protected by same policy |
| Zoom AI Companion | Meeting summaries, chat compose | Zoom states: Customer content (audio, video, chat, files) is not used to train AI models. Uses third-party LLMs with no-training agreements. | N/A – not trained on user data | ✓ Yes – commercial terms exclude training; admins can disable AI features if desired |
Across platforms, when you opt out of AI training, it typically applies only to future data. Content already collected and potentially already used for training stays in the models.
Implications:
LinkedIn's "private messages" and other platforms' messaging illustrate that "private" often means "only visible to participants," but platforms still reserve the right to analyze them for recommendations, AI, or safety.
Takeaway: Don't rely on labels. Read the data-use sections of privacy policies to understand what "private" actually protects against.
Almost every platform that offers enterprise or API tiers excludes training on customer data in those contracts.
Best practice: For any tool that handles confidential work, use the paid/enterprise version with a DPA or BAA, not the free consumer tier.
A clear pattern emerges: most platforms treat consumer/free users as potential training sources, while enterprise customers get contractual protection.
Typical terms:
Examples: Free ChatGPT, Claude.ai Free/Pro, consumer Gmail, LinkedIn personal accounts
Risk: Company work done on personal/free accounts may already be in training sets
Typical terms:
Examples: Slack (all paid plans), Claude API/Teams, ChatGPT Enterprise, Google Workspace, Microsoft 365
Protection: Strong, but verify contract language and periodically audit
Your company pays for Slack Enterprise and Google Workspace (protected). But employees also use:
These personal/free accounts don't have enterprise protections. Company data is leaking into consumer-tier training sets because employees don't realize the distinction.
Solution: Educate teams, provide approved enterprise tools, and ban use of consumer AI for business work in your acceptable use policy.
| Tool Category | Approved Tools (Enterprise Tier) | Required Settings / Contract Terms | Prohibited for Business Use |
|---|---|---|---|
| Team Chat | Slack (paid plans), Microsoft Teams | • Slack: "No LLM training" per 2025 policy • Teams: Copilot disabled or enterprise DPA in place |
Consumer messaging apps (WhatsApp, Telegram for business without enterprise accounts) |
| AI Assistants | ChatGPT Enterprise, Claude API/Teams, Copilot (M365) | • DPA excluding training on customer data • Admin controls to audit usage |
Free ChatGPT, free Claude.ai, consumer Gemini, Perplexity (for client work) |
| Code Repos | GitHub Enterprise, GitLab (paid) | • Private repos not used for Copilot training • Telemetry opt-out configured |
Public repos for proprietary code; free GitHub without Enterprise terms |
| Docs & Knowledge | Google Workspace, Notion (paid), Confluence | • Workspace: training excluded per terms • Notion: "no training" policy verified |
Consumer Google Docs (free Gmail), public Notion pages for sensitive content |
| Freelance Platforms | Upwork (with AI opt-outs configured) | • AI Preferences: opt out of work product + communications • Freelancer confirms same settings |
PeoplePerHour (messages "not confidential"), Freelancer.com (non-personal UGC) for confidential work |
| Professional Networks | LinkedIn (with AI toggle off) | • "Data for Generative AI Improvement" disabled • Confidential client discussions banned in LI messages |
Using LinkedIn messages for strategy, client names, or deal terms |
Your Upwork AI policy is just one piece. A comprehensive company policy should address:
Your "Upwork policy" should be consistent with your "Slack policy" and your "Claude policy."
Example inconsistency to avoid:
Result: Data you protected on one platform leaks through others.
Better approach:
This creates a consistent data perimeter.
Most companies discover their AI training exposure after they've already used dozens of tools with inadequate protections. By then, confidential data may already be in training sets—and retroactive opt-outs won't remove it.
I help businesses build comprehensive, enforceable AI and SaaS data policies before exposure occurs—and remediate exposure when it's already happened.
AI data training policies are evolving monthly. Generic business attorneys often lack the specific knowledge required:
Whether you're building an AI policy from scratch, responding to an exposure incident, or negotiating with vendors, I provide practical, enforceable solutions that protect your data across all platforms.
Send me your current tool stack list, any vendor contracts you have concerns about, and a summary of what types of data you need to protect. I'll evaluate your exposure and outline a policy and remediation strategy.
Email: owner@terms.law
AI policy audits: $400-$900. Contract drafting: ~$450 (typically 2 hours @ $240/hr). Vendor negotiation: $240/hr. Incident response: hourly or contingency arrangements available.