Mistral AI · Mistral AI Privacy Policy

Third-Party Training Dataset Disclosure

Medium severity
Share 𝕏 Share in Share 🔒 PDF

What it is

Mistral AI trains its AI models using datasets from third parties and publicly available internet data, which may contain your personal information even after filtering attempts.

Consumer impact (what this means for users)

Personal data about you that exists publicly online — such as on social media, news articles, or public records — may have been used to train Mistral AI's models, and there is no guarantee it was successfully filtered out despite stated efforts.

What you can do

⚠️ These actions may provide transparency or partial mitigation but may not fully address the underlying issue. Effectiveness varies by jurisdiction and individual circumstances.
  • Delete Your Data
    Submit a data erasure request via the Privacy Requests contact form at https://mistral.ai/en/contact, addressed to the DPO, identifying your personal data and requesting its removal from training datasets. Note that erasure from trained model weights may be technically complex and Mistral AI may respond with alternative remediation.

Cross-platform context

See how other platforms handle Third-Party Training Dataset Disclosure and similar clauses.

Compare across platforms →
Need full compliance memos? See Professional →

Why it matters (compliance & risk perspective)

Your personal data may have been included in AI training datasets scraped from the internet without your knowledge, and 'good practices to filter' does not guarantee removal.

View original clause language
Training Datasets. In some cases, we access datasets provided by third parties for our model training purposes. These datasets may include personal data (even if such third parties and Mistral AI use good practices to filter out such personal data), proprietary data, or public data. [...] Data publicly available on the Internet. Our artificial intelligence models are trained on data that is publicly available on the Internet by third parties, which may contain personal data, even if we use good practices to filter out such personal data.

Institutional analysis (Compliance & legal intelligence)

(1) REGULATORY FRAMEWORK: This provision implicates GDPR Art. 14 (transparency obligations for data collected from third parties), Art. 6 (lawful basis for training data processing), and Art. 17 (right to erasure from training datasets — a practically complex right). The EU AI Act (Regulation 2024/1689) Art. 53 imposes specific transparency and documentation obligations on general-purpose AI model providers regarding training data, including copyright and personal data governance documentation. The CNIL's 2024 framework on AI and personal data is directly applicable. (2)

🔒

Compliance intelligence locked

Regulatory citations, enforcement risk, and due diligence action items.

Watcher $9.99/mo Professional $149/mo

Watcher: regulatory citations. Professional: full compliance memo.

Applicable agencies

  • FTC
    The FTC has authority over unfair or deceptive practices involving the collection and use of consumer data for AI model training without adequate consent, including data scraped from public internet sources.
    File a complaint →

Provision details

Document information
Document
Mistral AI Privacy Policy
Entity
Mistral AI
Document last updated
April 29, 2026
Tracking information
First tracked
April 30, 2026
Last verified
April 30, 2026
Record ID
CA-P-004355
Document ID
CA-D-00443
Evidence Provenance
Source URL
Wayback Machine
SHA-256
73a02ec10fcf1627015be32bbcec27aa65278073cf29aaf0a9823340b9de2a08
Verified
✓ Snapshot stored   ✓ Change verified
How to Cite
ConductAtlas Policy Archive
Entity: Mistral AI | Document: Mistral AI Privacy Policy | Record: CA-P-004355
Captured: 2026-04-30 08:58:00 UTC | SHA-256: 73a02ec10fcf1627…
URL: https://conductatlas.com/platform/mistral-ai/mistral-ai-privacy-policy/third-party-training-dataset-disclosure/
Accessed: May 2, 2026
Classification
Severity
Medium
Categories

Other provisions in this document