Terug naar woordenlijst Beveiliging

Datavergiftiging

Aanvallen die opzettelijk AI-trainingsdata vergiftigen om kwetsbaarheden te creëren of modelgedrag te vervormen.

Understanding Data Poisoning

Data poisoning is an attack technique where adversaries deliberately introduce corrupted, mislabeled, or malicious samples into an AI system's training dataset. Because machine learning models learn patterns directly from data, even a small percentage of poisoned samples can significantly alter model behavior. These attacks are particularly dangerous because they occur before training, making the resulting vulnerabilities inherent to the model and difficult to detect through standard evaluation procedures.

Common Attack Patterns

Label-flipping attacks change the labels of training examples to cause systematic misclassification. Clean-label attacks insert correctly labeled but adversarially crafted examples that shift decision boundaries. Availability attacks degrade overall model performance by injecting noise. Targeted attacks cause specific misclassifications while maintaining general accuracy, making them harder to detect. Web-scraped datasets and crowd-sourced annotations are especially vulnerable, as attackers can contribute poisoned data through legitimate-appearing channels.

Enterprise Mitigation

Defending against data poisoning requires rigorous data governance throughout the AI lifecycle. Data provenance tracking ensures every training sample has a verified source. Statistical outlier detection and dataset sanitization techniques can identify anomalous samples before training. Training with robust loss functions reduces sensitivity to corrupted data points. For critical applications, enterprises should maintain curated, access-controlled datasets with versioning, conduct regular data audits, and implement anomaly detection on incoming data streams to catch poisoning attempts early.

Gerelateerde diensten en producten