Your AI Trains on Unmasked Data.

The EU AI Act requires documented data governance for high-risk AI. Agentforce training data needs masking before it touches your models.

MAXIMUM REGULATORY FINE

€35M

or 7% of global annual turnover: maximum EU AI Act fine for prohibited practices

Salesforce AI Without Training Data Governance

€35M

maximum fine for using prohibited AI practices: or 7% of global annual turnover

The EU AI Act (in force August 2024, phased enforcement 2025-2027) imposes strict requirements on high-risk AI systems. Including those trained on personal data. For Salesforce organizations using AI features like Agentforce, Einstein, or custom Apex ML models, training data governance is now a compliance obligation, not just a best practice.

100%

of training datasets for high-risk AI systems must be governed. Including data quality, relevance, and personal data minimization

EU AI Act Article 10 requires that training, validation, and testing datasets for high-risk AI systems undergo data governance practices: relevant, representative, free of errors, and handled in compliance with GDPR. For Salesforce teams, this means the customer data used to train or fine-tune AI models must be masked or anonymized before use.

2027

full EU AI Act enforcement deadline for general-purpose AI systems and high-risk AI categories

The EU AI Act has a phased enforcement timeline: prohibited AI practices (Feb 2025), codes of practice (Aug 2025), high-risk AI obligations (Aug 2026), and general-purpose AI obligations (Aug 2027). Organizations using Salesforce AI for customer scoring, service routing, or predictive analytics should classify their systems now and begin training data governance.

High-Risk AI Systems Must Govern Training Data

The EU AI Act (in force August 2024, phased enforcement 2025-2027) imposes strict requirements on high-risk AI systems. Including those trained on personal data. Three articles create specific obligations in Salesforce AI deployments:

Article 10

Training Data Quality & Governance

Training, validation, and testing datasets for high-risk AI systems must be governed for relevance, representativeness, and personal data minimization. Customer records used to train Agentforce or Einstein models must be masked or anonymized before use. Replacing real PII with realistic synthetic data.

DataMasker

Article 11

Technical Documentation of Training Data

High-risk AI providers must maintain technical documentation including training data sources, governance practices, and data quality measures. Regulators require timestamped records of: which records were included, which fields were masked, masking rules applied, and when the job ran.

DataMasker Audit Logs

Article 99

Enforcement & Penalties

Article 99 establishes penalties for prohibited AI practices and violations of high-risk AI obligations: up to €35M or 7% of global annual turnover. Regulators assess fines based on revenue, severity of violation, and cooperation during investigation.

Privacy Rights Automation

Training Data Masking, Retention, & Audit Compliance

Article 10: Training Data Quality

DataMasker

Mask PII Before AI Training Data Extraction

DataMasker masks Salesforce records before they are used as AI training data. Names, email addresses, phone numbers, and other personal identifiers are replaced with realistic synthetic data that preserves the statistical patterns AI models need. Without containing real PII. The masked dataset is GDPR-compliant and meets EU AI Act Article 10 data quality requirements.

Article 10: Data Minimization

Data Retention Manager

Enforce Training Data Retention Limits

EU AI Act and GDPR together require that training data is not retained longer than necessary. Data Retention Manager implements retention schedules for AI training datasets stored in Salesforce. When a model is deprecated or retrained, the associated training data is automatically purged according to your configured retention policy.

Article 11: Technical Documentation

Masking Audit Logs

Document Training Data Provenance for Audits

EU AI Act requires high-risk AI providers to maintain technical documentation including training data sources and governance practices. DataMasker's masking logs provide a complete audit trail: which records were included, which fields were masked, when the masking job ran, and which rules were applied. This documentation satisfies both EU AI Act technical documentation requirements and GDPR data processing records.

Key Takeaways

EU AI Act Article 10 data quality requirements addressed through data minimization before AI training

Agentforce and Einstein models should not reason over PII your org was never supposed to retain

Sandbox masking prevents AI models in development from training on live production personal data

Data Retention Manager removes obsolete records that would otherwise be in AI model training scope

High-risk AI system documentation supported by automated data inventory from Personal Data Discovery

Act applies from August 2026, organizations deploying Salesforce AI need data governance in place now

Frequently Asked Questions

EU AI Act Compliance Starts With Training Data Governance.

See how Salesforce organizations implement EU AI Act-compliant training data masking before Agentforce go-live.

100% native to Salesforce. Your data never leaves your org.